[ 520.863127] env[62753]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'linux_bridge' {{(pid=62753) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 520.863481] env[62753]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'noop' {{(pid=62753) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 520.863597] env[62753]: DEBUG os_vif [-] Loaded VIF plugin class '' with name 'ovs' {{(pid=62753) initialize /opt/stack/data/venv/lib/python3.10/site-packages/os_vif/__init__.py:44}} [ 520.863884] env[62753]: INFO os_vif [-] Loaded VIF plugins: linux_bridge, noop, ovs [ 520.958333] env[62753]: DEBUG oslo_concurrency.processutils [-] Running cmd (subprocess): grep -F node.session.scan /sbin/iscsiadm {{(pid=62753) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:390}} [ 520.968363] env[62753]: DEBUG oslo_concurrency.processutils [-] CMD "grep -F node.session.scan /sbin/iscsiadm" returned: 0 in 0.010s {{(pid=62753) execute /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/processutils.py:428}} [ 521.573765] env[62753]: INFO nova.virt.driver [None req-0a9eb144-8970-4855-bb03-c44432612441 None None] Loading compute driver 'vmwareapi.VMwareVCDriver' [ 521.644321] env[62753]: DEBUG oslo_concurrency.lockutils [-] Acquiring lock "oslo_vmware_api_lock" by "oslo_vmware.api.VMwareAPISession._create_session" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 521.644498] env[62753]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" acquired by "oslo_vmware.api.VMwareAPISession._create_session" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 521.644575] env[62753]: DEBUG oslo_vmware.service [-] Creating suds client with soap_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk' and wsdl_url='https://vc1.osci.c.eu-de-1.cloud.sap:443/sdk/vimService.wsdl' {{(pid=62753) __init__ /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:242}} [ 524.894423] env[62753]: DEBUG oslo_vmware.service [-] Invoking ServiceInstance.RetrieveServiceContent with opID=oslo.vmware-a7fc9739-633d-4c0f-81ad-106ac6efef4b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.910373] env[62753]: DEBUG oslo_vmware.api [-] Logging into host: vc1.osci.c.eu-de-1.cloud.sap. {{(pid=62753) _create_session /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:242}} [ 524.910515] env[62753]: DEBUG oslo_vmware.service [-] Invoking SessionManager.Login with opID=oslo.vmware-4dbd2ba4-58d4-4da6-8d59-25352e7c7279 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.950473] env[62753]: INFO oslo_vmware.api [-] Successfully established new session; session ID is 05f33. [ 524.950630] env[62753]: DEBUG oslo_concurrency.lockutils [-] Lock "oslo_vmware_api_lock" "released" by "oslo_vmware.api.VMwareAPISession._create_session" :: held 3.306s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 524.951214] env[62753]: INFO nova.virt.vmwareapi.driver [None req-0a9eb144-8970-4855-bb03-c44432612441 None None] VMware vCenter version: 7.0.3 [ 524.954671] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3470fb40-890b-483a-b282-17e5634ed58f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.972182] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-692b878e-cde8-412c-98c0-d8bddd3b081d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.978105] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2cf3b689-49d9-45df-9709-5bf06185e959 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.984729] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e485fa4c-a5ab-4c3d-a7ea-60e614a3ed2a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 524.997497] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37837471-058f-4921-a061-80cad1fdd6b7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.003298] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ef8d0f7-7cb5-48cc-9a7c-9b0c7b10dcf2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.033070] env[62753]: DEBUG oslo_vmware.service [-] Invoking ExtensionManager.FindExtension with opID=oslo.vmware-527d43f0-657b-45b7-9d2f-2b6e34b2b15d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 525.037894] env[62753]: DEBUG nova.virt.vmwareapi.driver [None req-0a9eb144-8970-4855-bb03-c44432612441 None None] Extension org.openstack.compute already exists. {{(pid=62753) _register_openstack_extension /opt/stack/nova/nova/virt/vmwareapi/driver.py:226}} [ 525.040613] env[62753]: INFO nova.compute.provider_config [None req-0a9eb144-8970-4855-bb03-c44432612441 None None] No provider configs found in /etc/nova/provider_config/. If files are present, ensure the Nova process has access. [ 525.544115] env[62753]: DEBUG nova.context [None req-0a9eb144-8970-4855-bb03-c44432612441 None None] Found 2 cells: 00000000-0000-0000-0000-000000000000(cell0),0803355c-018d-4ae6-83cd-b4f84c418e2d(cell1) {{(pid=62753) load_cells /opt/stack/nova/nova/context.py:464}} [ 525.546389] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] Acquiring lock "00000000-0000-0000-0000-000000000000" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.546617] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] Lock "00000000-0000-0000-0000-000000000000" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.547370] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] Lock "00000000-0000-0000-0000-000000000000" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.547791] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] Acquiring lock "0803355c-018d-4ae6-83cd-b4f84c418e2d" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 525.547974] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] Lock "0803355c-018d-4ae6-83cd-b4f84c418e2d" acquired by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 525.548983] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] Lock "0803355c-018d-4ae6-83cd-b4f84c418e2d" "released" by "nova.context.set_target_cell..get_or_set_cached_cell_and_set_connections" :: held 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 525.568807] env[62753]: INFO dbcounter [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] Registered counter for database nova_cell0 [ 525.576915] env[62753]: INFO dbcounter [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] Registered counter for database nova_cell1 [ 525.580328] env[62753]: DEBUG oslo_db.sqlalchemy.engines [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62753) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 525.580943] env[62753]: DEBUG oslo_db.sqlalchemy.engines [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] MySQL server mode set to STRICT_TRANS_TABLES,STRICT_ALL_TABLES,NO_ZERO_IN_DATE,NO_ZERO_DATE,ERROR_FOR_DIVISION_BY_ZERO,TRADITIONAL,NO_AUTO_CREATE_USER,NO_ENGINE_SUBSTITUTION {{(pid=62753) _check_effective_sql_mode /opt/stack/data/venv/lib/python3.10/site-packages/oslo_db/sqlalchemy/engines.py:342}} [ 525.585474] env[62753]: ERROR nova.db.main.api [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 525.585474] env[62753]: result = function(*args, **kwargs) [ 525.585474] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 525.585474] env[62753]: return func(*args, **kwargs) [ 525.585474] env[62753]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 525.585474] env[62753]: result = fn(*args, **kwargs) [ 525.585474] env[62753]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 525.585474] env[62753]: return f(*args, **kwargs) [ 525.585474] env[62753]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 525.585474] env[62753]: return db.service_get_minimum_version(context, binaries) [ 525.585474] env[62753]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 525.585474] env[62753]: _check_db_access() [ 525.585474] env[62753]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 525.585474] env[62753]: stacktrace = ''.join(traceback.format_stack()) [ 525.585474] env[62753]: [ 525.586460] env[62753]: ERROR nova.db.main.api [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] No DB access allowed in nova-compute: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 525.586460] env[62753]: result = function(*args, **kwargs) [ 525.586460] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 525.586460] env[62753]: return func(*args, **kwargs) [ 525.586460] env[62753]: File "/opt/stack/nova/nova/context.py", line 422, in gather_result [ 525.586460] env[62753]: result = fn(*args, **kwargs) [ 525.586460] env[62753]: File "/opt/stack/nova/nova/db/main/api.py", line 179, in wrapper [ 525.586460] env[62753]: return f(*args, **kwargs) [ 525.586460] env[62753]: File "/opt/stack/nova/nova/objects/service.py", line 554, in _db_service_get_minimum_version [ 525.586460] env[62753]: return db.service_get_minimum_version(context, binaries) [ 525.586460] env[62753]: File "/opt/stack/nova/nova/db/main/api.py", line 238, in wrapper [ 525.586460] env[62753]: _check_db_access() [ 525.586460] env[62753]: File "/opt/stack/nova/nova/db/main/api.py", line 188, in _check_db_access [ 525.586460] env[62753]: stacktrace = ''.join(traceback.format_stack()) [ 525.586460] env[62753]: [ 525.586854] env[62753]: WARNING nova.objects.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] Failed to get minimum service version for cell 00000000-0000-0000-0000-000000000000 [ 525.586976] env[62753]: WARNING nova.objects.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] Failed to get minimum service version for cell 0803355c-018d-4ae6-83cd-b4f84c418e2d [ 525.587436] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] Acquiring lock "singleton_lock" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 525.587601] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] Acquired lock "singleton_lock" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 525.587843] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] Releasing lock "singleton_lock" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 525.588178] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] Full set of CONF: {{(pid=62753) _wait_for_exit_or_signal /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/service.py:363}} [ 525.588324] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ******************************************************************************** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2806}} [ 525.588453] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] Configuration options gathered from: {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2807}} [ 525.588593] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] command line args: ['--config-file', '/etc/nova/nova.conf', '--config-file', '/etc/nova/nova-cpu-common.conf', '--config-file', '/etc/nova/nova-cpu-1.conf'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2808}} [ 525.588788] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] config files: ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2809}} [ 525.588917] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ================================================================================ {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2811}} [ 525.589173] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] allow_resize_to_same_host = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.589352] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] arq_binding_timeout = 300 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.589489] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] backdoor_port = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.589617] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] backdoor_socket = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.589781] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] block_device_allocate_retries = 60 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.589939] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] block_device_allocate_retries_interval = 3 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.590151] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cert = self.pem {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.590331] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] compute_driver = vmwareapi.VMwareVCDriver {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.590505] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] compute_monitors = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.590671] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] config_dir = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.590841] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] config_drive_format = iso9660 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.590973] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] config_file = ['/etc/nova/nova.conf', '/etc/nova/nova-cpu-common.conf', '/etc/nova/nova-cpu-1.conf'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.591171] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] config_source = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.591319] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] console_host = devstack {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.591504] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] control_exchange = nova {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.591676] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cpu_allocation_ratio = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.591836] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] daemon = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.592008] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] debug = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.592173] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] default_access_ip_network_name = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.592339] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] default_availability_zone = nova {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.592492] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] default_ephemeral_format = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.592649] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] default_green_pool_size = 1000 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.592880] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] default_log_levels = ['amqp=WARN', 'amqplib=WARN', 'boto=WARN', 'qpid=WARN', 'sqlalchemy=WARN', 'suds=INFO', 'oslo.messaging=INFO', 'oslo_messaging=INFO', 'iso8601=WARN', 'requests.packages.urllib3.connectionpool=WARN', 'urllib3.connectionpool=WARN', 'websocket=WARN', 'requests.packages.urllib3.util.retry=WARN', 'urllib3.util.retry=WARN', 'keystonemiddleware=WARN', 'routes.middleware=WARN', 'stevedore=WARN', 'taskflow=WARN', 'keystoneauth=WARN', 'oslo.cache=INFO', 'oslo_policy=INFO', 'dogpile.core.dogpile=INFO', 'glanceclient=WARN', 'oslo.privsep.daemon=INFO'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.593056] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] default_schedule_zone = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.593245] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] disk_allocation_ratio = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.593417] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] enable_new_services = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.593590] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] enabled_apis = ['osapi_compute'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.593752] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] enabled_ssl_apis = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.593910] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] flat_injected = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.594076] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] force_config_drive = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.594238] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] force_raw_images = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.594406] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] graceful_shutdown_timeout = 5 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.594563] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] heal_instance_info_cache_interval = 60 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.594773] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] host = cpu-1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.594947] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] initial_cpu_allocation_ratio = 4.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.595122] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] initial_disk_allocation_ratio = 1.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.595286] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] initial_ram_allocation_ratio = 1.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.595496] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] injected_network_template = /opt/stack/nova/nova/virt/interfaces.template {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.595659] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] instance_build_timeout = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.595818] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] instance_delete_interval = 300 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.595983] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] instance_format = [instance: %(uuid)s] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.596187] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] instance_name_template = instance-%08x {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.596367] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] instance_usage_audit = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.596541] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] instance_usage_audit_period = month {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.596706] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] instance_uuid_format = [instance: %(uuid)s] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.596868] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] instances_path = /opt/stack/data/nova/instances {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.597047] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] internal_service_availability_zone = internal {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.597211] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] key = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.597370] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] live_migration_retry_count = 30 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.597537] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] log_color = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.597699] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] log_config_append = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.597864] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] log_date_format = %Y-%m-%d %H:%M:%S {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.598030] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] log_dir = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.598195] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] log_file = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.598326] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] log_options = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.598486] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] log_rotate_interval = 1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.598652] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] log_rotate_interval_type = days {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.598817] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] log_rotation_type = none {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.598945] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] logging_context_format_string = %(color)s%(levelname)s %(name)s [%(global_request_id)s %(request_id)s %(project_name)s %(user_name)s%(color)s] %(instance)s%(color)s%(message)s {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.599081] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] logging_debug_format_suffix = {{(pid=%(process)d) %(funcName)s %(pathname)s:%(lineno)d}} {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.599286] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] logging_default_format_string = %(color)s%(levelname)s %(name)s [-%(color)s] %(instance)s%(color)s%(message)s {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.599465] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] logging_exception_prefix = ERROR %(name)s %(instance)s {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.599595] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] logging_user_identity_format = %(user)s %(project)s %(domain)s %(system_scope)s %(user_domain)s %(project_domain)s {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.599757] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] long_rpc_timeout = 1800 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.599917] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] max_concurrent_builds = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.600088] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] max_concurrent_live_migrations = 1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.600257] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] max_concurrent_snapshots = 5 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.600414] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] max_local_block_devices = 3 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.600570] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] max_logfile_count = 30 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.600726] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] max_logfile_size_mb = 200 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.600883] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] maximum_instance_delete_attempts = 5 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.601058] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] metadata_listen = 0.0.0.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.601270] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] metadata_listen_port = 8775 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.601396] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] metadata_workers = 2 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.601574] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] migrate_max_retries = -1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.601749] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] mkisofs_cmd = genisoimage {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.601956] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] my_block_storage_ip = 10.180.1.21 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.602106] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] my_ip = 10.180.1.21 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.602297] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] network_allocate_retries = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.602487] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] non_inheritable_image_properties = ['cache_in_nova', 'bittorrent'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.602657] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] osapi_compute_listen = 0.0.0.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.602823] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] osapi_compute_listen_port = 8774 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.602990] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] osapi_compute_unique_server_name_scope = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.603172] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] osapi_compute_workers = 2 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.603337] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] password_length = 12 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.603497] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] periodic_enable = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.603657] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] periodic_fuzzy_delay = 60 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.603826] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] pointer_model = usbtablet {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.603992] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] preallocate_images = none {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.604166] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] publish_errors = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.604298] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] pybasedir = /opt/stack/nova {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.604456] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ram_allocation_ratio = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.604614] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] rate_limit_burst = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.604780] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] rate_limit_except_level = CRITICAL {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.604937] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] rate_limit_interval = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.605108] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] reboot_timeout = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.605291] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] reclaim_instance_interval = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.605462] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] record = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.605632] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] reimage_timeout_per_gb = 60 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.605796] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] report_interval = 120 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.605956] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] rescue_timeout = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.606128] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] reserved_host_cpus = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.606290] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] reserved_host_disk_mb = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.606448] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] reserved_host_memory_mb = 512 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.606605] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] reserved_huge_pages = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.606764] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] resize_confirm_window = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.606922] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] resize_fs_using_block_device = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.607088] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] resume_guests_state_on_host_boot = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.607266] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] rootwrap_config = /etc/nova/rootwrap.conf {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.607437] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] rpc_response_timeout = 60 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.607596] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] run_external_periodic_tasks = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.607762] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] running_deleted_instance_action = reap {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.607917] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] running_deleted_instance_poll_interval = 1800 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.608084] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] running_deleted_instance_timeout = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.608252] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] scheduler_instance_sync_interval = 120 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.608440] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] service_down_time = 720 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.608609] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] servicegroup_driver = db {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.608763] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] shell_completion = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.608924] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] shelved_offload_time = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.609093] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] shelved_poll_interval = 3600 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.609291] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] shutdown_timeout = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.609458] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] source_is_ipv6 = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.609615] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ssl_only = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.609857] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] state_path = /opt/stack/data/n-cpu-1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.610034] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] sync_power_state_interval = 600 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.610200] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] sync_power_state_pool_size = 1000 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.610368] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] syslog_log_facility = LOG_USER {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.610525] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] tempdir = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.610683] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] timeout_nbd = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.610846] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] transport_url = **** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.611009] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] update_resources_interval = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.611174] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] use_cow_images = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.611367] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] use_eventlog = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.611546] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] use_journal = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.611719] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] use_json = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.611880] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] use_rootwrap_daemon = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.612049] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] use_stderr = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.612216] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] use_syslog = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.612373] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vcpu_pin_set = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.612541] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vif_plugging_is_fatal = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.612706] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vif_plugging_timeout = 300 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.612872] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] virt_mkfs = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.613042] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] volume_usage_poll_interval = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.613209] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] watch_log_file = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.613377] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] web = /usr/share/spice-html5 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2819}} [ 525.613561] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_brick.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.613725] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_brick.wait_mpath_device_attempts = 4 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.613885] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_brick.wait_mpath_device_interval = 1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.614066] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_concurrency.disable_process_locking = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.614660] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_concurrency.lock_path = /opt/stack/data/n-cpu-1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.614859] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_metrics.metrics_buffer_size = 1000 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.615045] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_metrics.metrics_enabled = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.615231] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_metrics.metrics_process_name = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.615409] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_metrics.metrics_socket_file = /var/tmp/metrics_collector.sock {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.615578] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_metrics.metrics_thread_stop_timeout = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.615763] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.auth_strategy = keystone {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.615932] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.compute_link_prefix = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.616121] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.config_drive_skip_versions = 1.0 2007-01-19 2007-03-01 2007-08-29 2007-10-10 2007-12-15 2008-02-01 2008-09-01 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.616300] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.dhcp_domain = novalocal {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.616472] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.enable_instance_password = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.616641] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.glance_link_prefix = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.616810] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.instance_list_cells_batch_fixed_size = 100 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.616982] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.instance_list_cells_batch_strategy = distributed {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.617193] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.instance_list_per_project_cells = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.617448] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.list_records_by_skipping_down_cells = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.617644] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.local_metadata_per_cell = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.617822] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.max_limit = 1000 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.617994] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.metadata_cache_expiration = 15 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.618186] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.neutron_default_tenant_id = default {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.618360] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.response_validation = warn {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.618533] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.use_neutron_default_nets = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.618713] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.vendordata_dynamic_connect_timeout = 5 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.618880] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.vendordata_dynamic_failure_fatal = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.619060] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.vendordata_dynamic_read_timeout = 5 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.619307] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.vendordata_dynamic_ssl_certfile = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.619588] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.vendordata_dynamic_targets = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.619802] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.vendordata_jsonfile_path = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.620048] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api.vendordata_providers = ['StaticJSON'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.620349] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.backend = dogpile.cache.memcached {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.620568] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.backend_argument = **** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.620783] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.config_prefix = cache.oslo {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.620986] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.dead_timeout = 60.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.621203] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.debug_cache_backend = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.621461] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.enable_retry_client = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.621609] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.enable_socket_keepalive = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.621816] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.enabled = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.622015] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.enforce_fips_mode = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.622234] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.expiration_time = 600 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.622431] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.hashclient_retry_attempts = 2 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.622622] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.hashclient_retry_delay = 1.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.622836] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.memcache_dead_retry = 300 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.623033] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.memcache_password = **** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.623220] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.memcache_pool_connection_get_timeout = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.623390] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.memcache_pool_flush_on_reconnect = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.623555] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.memcache_pool_maxsize = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.623718] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.memcache_pool_unused_timeout = 60 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.623895] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.memcache_sasl_enabled = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.624127] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.memcache_servers = ['localhost:11211'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.624356] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.memcache_socket_timeout = 1.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.624544] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.memcache_username = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.624745] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.proxies = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.624954] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.redis_db = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.625178] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.redis_password = **** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.625392] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.redis_sentinel_service_name = mymaster {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.625609] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.redis_sentinels = ['localhost:26379'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.625838] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.redis_server = localhost:6379 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.626053] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.redis_socket_timeout = 1.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.626258] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.redis_username = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.626473] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.retry_attempts = 2 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.626685] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.retry_delay = 0.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.626893] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.socket_keepalive_count = 1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.627105] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.socket_keepalive_idle = 1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.627622] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.socket_keepalive_interval = 1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.627622] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.tls_allowed_ciphers = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.627737] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.tls_cafile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.627916] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.tls_certfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.628130] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.tls_enabled = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.628336] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cache.tls_keyfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.628544] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cinder.auth_section = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.628767] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cinder.auth_type = password {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.628956] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cinder.cafile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.629199] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cinder.catalog_info = volumev3::publicURL {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.629417] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cinder.certfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.629630] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cinder.collect_timing = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.629821] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cinder.cross_az_attach = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.630046] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cinder.debug = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.630280] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cinder.endpoint_template = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.630496] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cinder.http_retries = 3 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.630702] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cinder.insecure = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.630895] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cinder.keyfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.631105] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cinder.os_region_name = RegionOne {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.631310] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cinder.split_loggers = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.631495] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cinder.timeout = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.631707] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] compute.consecutive_build_service_disable_threshold = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.631901] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] compute.cpu_dedicated_set = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.632111] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] compute.cpu_shared_set = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.632295] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] compute.image_type_exclude_list = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.632487] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] compute.live_migration_wait_for_vif_plug = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.632700] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] compute.max_concurrent_disk_ops = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.632904] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] compute.max_disk_devices_to_attach = -1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.633108] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] compute.packing_host_numa_cells_allocation_strategy = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.633292] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] compute.provider_config_location = /etc/nova/provider_config/ {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.633463] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] compute.resource_provider_association_refresh = 300 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.633629] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] compute.sharing_providers_max_uuids_per_request = 200 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.633794] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] compute.shutdown_retry_interval = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.633977] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] compute.vmdk_allowed_types = ['streamOptimized', 'monolithicSparse'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.634175] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] conductor.workers = 2 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.634356] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] console.allowed_origins = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.634516] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] console.ssl_ciphers = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.634685] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] console.ssl_minimum_version = default {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.634855] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] consoleauth.enforce_session_timeout = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.635032] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] consoleauth.token_ttl = 600 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.635207] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.cafile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.635369] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.certfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.635533] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.collect_timing = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.635692] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.connect_retries = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.635851] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.connect_retry_delay = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.636015] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.endpoint_override = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.636187] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.insecure = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.636348] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.keyfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.636508] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.max_version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.636665] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.min_version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.636821] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.region_name = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.636979] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.retriable_status_codes = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.637167] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.service_name = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.637358] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.service_type = accelerator {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.637526] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.split_loggers = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.637686] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.status_code_retries = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.637858] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.status_code_retry_delay = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.638036] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.timeout = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.638253] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.valid_interfaces = ['internal', 'public'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.638424] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] cyborg.version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.638607] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.backend = sqlalchemy {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.638778] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.connection = **** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.638946] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.connection_debug = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.639131] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.connection_parameters = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.639326] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.connection_recycle_time = 3600 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.639496] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.connection_trace = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.639663] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.db_inc_retry_interval = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.639828] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.db_max_retries = 20 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.639992] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.db_max_retry_interval = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.640189] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.db_retry_interval = 1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.640368] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.max_overflow = 50 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.640534] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.max_pool_size = 5 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.640696] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.max_retries = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.640866] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.mysql_sql_mode = TRADITIONAL {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.641036] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.mysql_wsrep_sync_wait = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.641202] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.pool_timeout = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.641368] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.retry_interval = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.641529] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.slave_connection = **** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.641739] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.sqlite_synchronous = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.641864] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] database.use_db_reconnect = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.642019] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.backend = sqlalchemy {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.642195] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.connection = **** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.642363] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.connection_debug = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.642531] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.connection_parameters = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.642694] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.connection_recycle_time = 3600 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.642857] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.connection_trace = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.643028] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.db_inc_retry_interval = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.643196] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.db_max_retries = 20 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.643361] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.db_max_retry_interval = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.643522] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.db_retry_interval = 1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.643682] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.max_overflow = 50 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.643843] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.max_pool_size = 5 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.644009] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.max_retries = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.644190] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.mysql_sql_mode = TRADITIONAL {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.644354] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.mysql_wsrep_sync_wait = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.644512] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.pool_timeout = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.644672] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.retry_interval = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.644830] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.slave_connection = **** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.644992] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] api_database.sqlite_synchronous = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.645184] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] devices.enabled_mdev_types = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.645366] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ephemeral_storage_encryption.cipher = aes-xts-plain64 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.645538] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ephemeral_storage_encryption.default_format = luks {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.645705] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ephemeral_storage_encryption.enabled = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.645867] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ephemeral_storage_encryption.key_size = 512 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.646047] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.api_servers = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.646215] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.cafile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.646379] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.certfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.646542] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.collect_timing = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.646702] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.connect_retries = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.646860] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.connect_retry_delay = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.647029] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.debug = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.647228] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.default_trusted_certificate_ids = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.647407] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.enable_certificate_validation = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.647574] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.enable_rbd_download = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.647730] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.endpoint_override = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.647895] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.insecure = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.648068] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.keyfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.648232] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.max_version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.648392] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.min_version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.648554] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.num_retries = 3 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.648722] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.rbd_ceph_conf = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.648882] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.rbd_connect_timeout = 5 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.649059] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.rbd_pool = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.649261] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.rbd_user = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.649421] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.region_name = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.649584] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.retriable_status_codes = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.649740] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.service_name = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.649910] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.service_type = image {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.650106] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.split_loggers = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.650310] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.status_code_retries = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.650479] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.status_code_retry_delay = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.650639] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.timeout = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.650819] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.valid_interfaces = ['internal', 'public'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.650985] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.verify_glance_signatures = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.651160] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] glance.version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.651331] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] guestfs.debug = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.651499] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] mks.enabled = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.651860] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] mks.mksproxy_base_url = http://127.0.0.1:6090/ {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.652063] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] image_cache.manager_interval = 2400 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.652238] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] image_cache.precache_concurrency = 1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.652412] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] image_cache.remove_unused_base_images = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.652582] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] image_cache.remove_unused_original_minimum_age_seconds = 86400 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.652749] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] image_cache.remove_unused_resized_minimum_age_seconds = 3600 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.652923] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] image_cache.subdirectory_name = _base {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.653112] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.api_max_retries = 60 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.653282] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.api_retry_interval = 2 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.653446] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.auth_section = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.653607] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.auth_type = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.653766] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.cafile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.653924] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.certfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.654098] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.collect_timing = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.654262] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.conductor_group = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.654422] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.connect_retries = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.654579] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.connect_retry_delay = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.654734] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.endpoint_override = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.654894] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.insecure = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.655062] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.keyfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.655224] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.max_version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.655384] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.min_version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.655547] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.peer_list = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.655703] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.region_name = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.655861] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.retriable_status_codes = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.656030] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.serial_console_state_timeout = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.656195] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.service_name = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.656367] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.service_type = baremetal {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.656523] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.shard = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.656682] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.split_loggers = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.656838] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.status_code_retries = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.656994] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.status_code_retry_delay = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.657187] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.timeout = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.657386] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.valid_interfaces = ['internal', 'public'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.657559] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ironic.version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.657734] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] key_manager.backend = nova.keymgr.conf_key_mgr.ConfKeyManager {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.657906] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] key_manager.fixed_key = **** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.658099] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican.auth_endpoint = http://localhost/identity/v3 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.658266] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican.barbican_api_version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.658426] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican.barbican_endpoint = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.658594] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican.barbican_endpoint_type = public {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.658752] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican.barbican_region_name = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.658909] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican.cafile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.659074] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican.certfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.659267] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican.collect_timing = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.659438] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican.insecure = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.659596] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican.keyfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.659759] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican.number_of_retries = 60 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.659920] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican.retry_delay = 1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.660101] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican.send_service_user_token = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.660287] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican.split_loggers = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.660449] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican.timeout = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.660611] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican.verify_ssl = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.660766] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican.verify_ssl_path = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.660934] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican_service_user.auth_section = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.661109] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican_service_user.auth_type = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.661268] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican_service_user.cafile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.661428] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican_service_user.certfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.661593] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican_service_user.collect_timing = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.661751] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican_service_user.insecure = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.661940] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican_service_user.keyfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.662080] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican_service_user.split_loggers = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.662242] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] barbican_service_user.timeout = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.662410] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vault.approle_role_id = **** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.662569] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vault.approle_secret_id = **** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.662735] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vault.kv_mountpoint = secret {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.662897] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vault.kv_path = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.663070] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vault.kv_version = 2 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.663239] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vault.namespace = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.663401] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vault.root_token_id = **** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.663557] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vault.ssl_ca_crt_file = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.663723] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vault.timeout = 60.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.663886] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vault.use_ssl = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.664062] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vault.vault_url = http://127.0.0.1:8200 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.664237] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.auth_section = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.664406] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.auth_type = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.664563] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.cafile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.664719] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.certfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.664880] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.collect_timing = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.665050] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.connect_retries = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.665215] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.connect_retry_delay = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.665377] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.endpoint_override = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.665539] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.insecure = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.665696] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.keyfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.665854] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.max_version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.666015] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.min_version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.666177] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.region_name = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.666336] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.retriable_status_codes = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.666491] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.service_name = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.666657] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.service_type = identity {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.666819] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.split_loggers = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.666975] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.status_code_retries = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.667146] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.status_code_retry_delay = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.667308] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.timeout = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.667486] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.valid_interfaces = ['internal', 'public'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.667647] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] keystone.version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.667840] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.connection_uri = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.668007] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.cpu_mode = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.668177] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.cpu_model_extra_flags = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.668348] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.cpu_models = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.668517] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.cpu_power_governor_high = performance {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.668683] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.cpu_power_governor_low = powersave {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.668844] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.cpu_power_management = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.669024] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.cpu_power_management_strategy = cpu_state {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.669209] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.device_detach_attempts = 8 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.669381] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.device_detach_timeout = 20 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.669547] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.disk_cachemodes = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.669706] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.disk_prefix = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.669871] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.enabled_perf_events = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.670046] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.file_backed_memory = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.670241] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.gid_maps = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.670406] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.hw_disk_discard = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.670564] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.hw_machine_type = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.670732] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.images_rbd_ceph_conf = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.670900] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.images_rbd_glance_copy_poll_interval = 15 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.671075] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.images_rbd_glance_copy_timeout = 600 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.671248] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.images_rbd_glance_store_name = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.671420] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.images_rbd_pool = rbd {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.671590] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.images_type = default {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.671746] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.images_volume_group = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.671905] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.inject_key = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.672079] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.inject_partition = -2 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.672242] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.inject_password = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.672409] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.iscsi_iface = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.672573] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.iser_use_multipath = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.672738] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.live_migration_bandwidth = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.672902] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.live_migration_completion_timeout = 800 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.673076] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.live_migration_downtime = 500 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.673245] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.live_migration_downtime_delay = 75 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.673411] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.live_migration_downtime_steps = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.673572] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.live_migration_inbound_addr = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.673733] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.live_migration_permit_auto_converge = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.673898] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.live_migration_permit_post_copy = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.674072] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.live_migration_scheme = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.674250] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.live_migration_timeout_action = abort {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.674417] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.live_migration_tunnelled = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.674577] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.live_migration_uri = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.674738] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.live_migration_with_native_tls = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.674897] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.max_queues = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.675070] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.mem_stats_period_seconds = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.675318] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.migration_inbound_addr = 10.180.1.21 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.675488] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.nfs_mount_options = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.675799] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.nfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.675974] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.num_aoe_discover_tries = 3 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.676154] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.num_iser_scan_tries = 5 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.676320] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.num_memory_encrypted_guests = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.676485] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.num_nvme_discover_tries = 5 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.676651] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.num_pcie_ports = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.676822] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.num_volume_scan_tries = 5 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.676990] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.pmem_namespaces = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.677167] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.quobyte_client_cfg = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.677459] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.quobyte_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.677642] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.rbd_connect_timeout = 5 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.677803] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.rbd_destroy_volume_retries = 12 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.677969] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.rbd_destroy_volume_retry_interval = 5 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.678144] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.rbd_secret_uuid = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.678313] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.rbd_user = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.678479] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.realtime_scheduler_priority = 1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.678649] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.remote_filesystem_transport = ssh {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.678810] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.rescue_image_id = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.678970] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.rescue_kernel_id = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.679158] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.rescue_ramdisk_id = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.679346] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.rng_dev_path = /dev/urandom {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.679510] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.rx_queue_size = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.679676] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.smbfs_mount_options = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.679952] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.smbfs_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.680152] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.snapshot_compression = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.680341] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.snapshot_image_format = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.680562] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.snapshots_directory = /opt/stack/data/nova/instances/snapshots {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.680732] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.sparse_logical_volumes = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.680895] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.swtpm_enabled = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.681072] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.swtpm_group = tss {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.681245] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.swtpm_user = tss {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.681417] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.sysinfo_serial = unique {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.681576] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.tb_cache_size = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.681734] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.tx_queue_size = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.681898] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.uid_maps = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.682074] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.use_virtio_for_bridges = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.682284] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.virt_type = kvm {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.682411] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.volume_clear = zero {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.682576] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.volume_clear_size = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.682740] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.volume_use_multipath = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.682898] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.vzstorage_cache_path = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.683075] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.vzstorage_log_path = /var/log/vstorage/%(cluster_name)s/nova.log.gz {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.683246] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.vzstorage_mount_group = qemu {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.683414] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.vzstorage_mount_opts = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.683581] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.vzstorage_mount_perms = 0770 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.683852] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.vzstorage_mount_point_base = /opt/stack/data/n-cpu-1/mnt {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.684038] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.vzstorage_mount_user = stack {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.684212] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] libvirt.wait_soft_reboot_seconds = 120 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.684390] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.auth_section = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.684564] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.auth_type = password {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.684727] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.cafile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.684887] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.certfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.685062] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.collect_timing = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.685229] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.connect_retries = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.685391] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.connect_retry_delay = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.685559] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.default_floating_pool = public {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.685719] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.endpoint_override = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.685884] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.extension_sync_interval = 600 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.686055] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.http_retries = 3 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.686224] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.insecure = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.686386] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.keyfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.686545] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.max_version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.686712] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.metadata_proxy_shared_secret = **** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.686873] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.min_version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.687059] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.ovs_bridge = br-int {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.687231] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.physnets = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.687404] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.region_name = RegionOne {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.687563] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.retriable_status_codes = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.687728] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.service_metadata_proxy = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.687884] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.service_name = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.688058] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.service_type = network {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.688226] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.split_loggers = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.688387] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.status_code_retries = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.688546] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.status_code_retry_delay = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.688702] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.timeout = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.688879] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.valid_interfaces = ['internal', 'public'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.689051] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] neutron.version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.689254] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] notifications.bdms_in_notifications = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.689444] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] notifications.default_level = INFO {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.689618] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] notifications.notification_format = unversioned {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.689781] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] notifications.notify_on_state_change = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.689957] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] notifications.versioned_notifications_topics = ['versioned_notifications'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.690161] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] pci.alias = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.690343] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] pci.device_spec = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.690512] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] pci.report_in_placement = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.690686] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.auth_section = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.690862] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.auth_type = password {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.691041] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.auth_url = http://10.180.1.21/identity {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.691207] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.cafile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.691366] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.certfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.691530] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.collect_timing = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.691689] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.connect_retries = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.691846] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.connect_retry_delay = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.692011] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.default_domain_id = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.692176] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.default_domain_name = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.692378] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.domain_id = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.692521] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.domain_name = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.692679] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.endpoint_override = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.692841] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.insecure = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.692997] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.keyfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.693170] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.max_version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.693330] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.min_version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.693495] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.password = **** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.693654] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.project_domain_id = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.693822] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.project_domain_name = Default {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.693991] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.project_id = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.694178] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.project_name = service {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.694351] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.region_name = RegionOne {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.694513] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.retriable_status_codes = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.694673] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.service_name = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.694841] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.service_type = placement {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.695009] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.split_loggers = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.695178] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.status_code_retries = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.695339] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.status_code_retry_delay = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.695497] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.system_scope = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.695654] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.timeout = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.695811] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.trust_id = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.695967] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.user_domain_id = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.696148] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.user_domain_name = Default {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.696314] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.user_id = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.696487] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.username = nova {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.696666] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.valid_interfaces = ['internal', 'public'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.696827] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] placement.version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.697010] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] quota.cores = 20 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.697184] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] quota.count_usage_from_placement = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.697357] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] quota.driver = nova.quota.DbQuotaDriver {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.697531] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] quota.injected_file_content_bytes = 10240 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.697694] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] quota.injected_file_path_length = 255 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.697859] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] quota.injected_files = 5 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.698032] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] quota.instances = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.698204] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] quota.key_pairs = 100 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.698369] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] quota.metadata_items = 128 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.698530] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] quota.ram = 51200 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.698690] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] quota.recheck_quota = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.698855] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] quota.server_group_members = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.699028] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] quota.server_groups = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.699212] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] scheduler.discover_hosts_in_cells_interval = -1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.699376] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] scheduler.enable_isolated_aggregate_filtering = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.699537] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] scheduler.image_metadata_prefilter = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.699695] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] scheduler.limit_tenants_to_placement_aggregate = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.699857] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] scheduler.max_attempts = 3 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.700026] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] scheduler.max_placement_results = 1000 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.700196] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] scheduler.placement_aggregate_required_for_tenants = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.700396] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] scheduler.query_placement_for_image_type_support = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.700564] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] scheduler.query_placement_for_routed_network_aggregates = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.700736] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] scheduler.workers = 2 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.700905] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.aggregate_image_properties_isolation_namespace = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.701087] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.aggregate_image_properties_isolation_separator = . {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.701268] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.available_filters = ['nova.scheduler.filters.all_filters'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.701436] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.build_failure_weight_multiplier = 1000000.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.701602] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.cpu_weight_multiplier = 1.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.701766] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.cross_cell_move_weight_multiplier = 1000000.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.701933] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.disk_weight_multiplier = 1.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.702137] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.enabled_filters = ['ComputeFilter', 'ComputeCapabilitiesFilter', 'ImagePropertiesFilter', 'ServerGroupAntiAffinityFilter', 'ServerGroupAffinityFilter', 'SameHostFilter', 'DifferentHostFilter'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.702312] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.host_subset_size = 1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.702482] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.hypervisor_version_weight_multiplier = 1.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.702639] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.image_properties_default_architecture = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.702802] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.io_ops_weight_multiplier = -1.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.702968] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.isolated_hosts = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.703145] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.isolated_images = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.703310] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.max_instances_per_host = 50 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.703468] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.max_io_ops_per_host = 8 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.703634] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.num_instances_weight_multiplier = 0.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.703795] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.pci_in_placement = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.703958] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.pci_weight_multiplier = 1.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.704132] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.ram_weight_multiplier = 1.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.704298] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.restrict_isolated_hosts_to_isolated_images = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.704457] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.shuffle_best_same_weighed_hosts = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.704619] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.soft_affinity_weight_multiplier = 1.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.704783] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.soft_anti_affinity_weight_multiplier = 1.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.704947] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.track_instance_changes = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.705137] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] filter_scheduler.weight_classes = ['nova.scheduler.weights.all_weighers'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.705318] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] metrics.required = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.705483] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] metrics.weight_multiplier = 1.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.705646] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] metrics.weight_of_unavailable = -10000.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.705810] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] metrics.weight_setting = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.706134] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] serial_console.base_url = ws://127.0.0.1:6083/ {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.706314] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] serial_console.enabled = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.706491] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] serial_console.port_range = 10000:20000 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.706662] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] serial_console.proxyclient_address = 127.0.0.1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.706830] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] serial_console.serialproxy_host = 0.0.0.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.706998] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] serial_console.serialproxy_port = 6083 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.707184] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] service_user.auth_section = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.707366] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] service_user.auth_type = password {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.707538] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] service_user.cafile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.707697] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] service_user.certfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.707858] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] service_user.collect_timing = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.708025] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] service_user.insecure = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.708189] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] service_user.keyfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.708362] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] service_user.send_service_user_token = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.708527] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] service_user.split_loggers = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.708700] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] service_user.timeout = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.708870] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] spice.agent_enabled = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.709044] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] spice.enabled = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.709386] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] spice.html5proxy_base_url = http://127.0.0.1:6082/spice_auto.html {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.709585] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] spice.html5proxy_host = 0.0.0.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.709758] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] spice.html5proxy_port = 6082 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.709922] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] spice.image_compression = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.710101] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] spice.jpeg_compression = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.710277] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] spice.playback_compression = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.710450] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] spice.require_secure = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.710623] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] spice.server_listen = 127.0.0.1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.710795] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] spice.server_proxyclient_address = 127.0.0.1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.710956] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] spice.streaming_mode = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.711129] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] spice.zlib_compression = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.711300] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] upgrade_levels.baseapi = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.711486] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] upgrade_levels.compute = auto {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.711677] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] upgrade_levels.conductor = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.711840] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] upgrade_levels.scheduler = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.712013] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vendordata_dynamic_auth.auth_section = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.712183] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vendordata_dynamic_auth.auth_type = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.712346] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vendordata_dynamic_auth.cafile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.712498] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vendordata_dynamic_auth.certfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.712664] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vendordata_dynamic_auth.collect_timing = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.712824] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vendordata_dynamic_auth.insecure = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.712981] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vendordata_dynamic_auth.keyfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.713158] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vendordata_dynamic_auth.split_loggers = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.713320] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vendordata_dynamic_auth.timeout = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.713494] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.api_retry_count = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.713653] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.ca_file = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.713823] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.cache_prefix = devstack-image-cache {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.713988] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.cluster_name = testcl1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.714167] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.connection_pool_size = 10 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.714330] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.console_delay_seconds = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.714497] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.datastore_regex = ^datastore.* {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.714699] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.host_ip = vc1.osci.c.eu-de-1.cloud.sap {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.714871] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.host_password = **** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.715047] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.host_port = 443 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.715221] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.host_username = administrator@vsphere.local {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.715393] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.insecure = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.715555] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.integration_bridge = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.715718] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.maximum_objects = 100 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.715877] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.pbm_default_policy = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.716049] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.pbm_enabled = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.716212] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.pbm_wsdl_location = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.716384] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.serial_log_dir = /opt/vmware/vspc {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.716543] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.serial_port_proxy_uri = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.716700] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.serial_port_service_uri = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.716868] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.task_poll_interval = 0.5 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.717049] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.use_linked_clone = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.717224] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.vnc_keymap = en-us {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.717391] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.vnc_port = 5900 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.717556] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vmware.vnc_port_total = 10000 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.717741] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vnc.auth_schemes = ['none'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.717913] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vnc.enabled = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.718218] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vnc.novncproxy_base_url = http://127.0.0.1:6080/vnc_auto.html {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.718407] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vnc.novncproxy_host = 0.0.0.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.718577] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vnc.novncproxy_port = 6080 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.718752] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vnc.server_listen = 127.0.0.1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.718924] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vnc.server_proxyclient_address = 127.0.0.1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.719096] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vnc.vencrypt_ca_certs = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.719289] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vnc.vencrypt_client_cert = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.719456] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vnc.vencrypt_client_key = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.719636] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.disable_compute_service_check_for_ffu = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.719801] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.disable_deep_image_inspection = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.719965] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.disable_fallback_pcpu_query = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.720163] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.disable_group_policy_check_upcall = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.720338] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.disable_libvirt_livesnapshot = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.720503] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.disable_rootwrap = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.720666] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.enable_numa_live_migration = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.720828] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.enable_qemu_monitor_announce_self = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.720992] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.ensure_libvirt_rbd_instance_dir_cleanup = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.721171] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.handle_virt_lifecycle_events = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.721337] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.libvirt_disable_apic = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.721525] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.never_download_image_if_on_rbd = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.721754] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.qemu_monitor_announce_self_count = 3 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.721936] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.qemu_monitor_announce_self_interval = 1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.722116] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.reserve_disk_resource_for_image_cache = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.722286] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.skip_cpu_compare_at_startup = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.722453] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.skip_cpu_compare_on_dest = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.722618] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.skip_hypervisor_version_check_on_lm = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.722781] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.skip_reserve_in_use_ironic_nodes = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.722941] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.unified_limits_count_pcpu_as_vcpu = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.723122] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] workarounds.wait_for_vif_plugged_event_during_hard_reboot = [] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.723308] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] wsgi.api_paste_config = /etc/nova/api-paste.ini {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.723481] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] wsgi.client_socket_timeout = 900 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.723648] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] wsgi.default_pool_size = 1000 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.723816] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] wsgi.keep_alive = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.723984] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] wsgi.max_header_line = 16384 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.724161] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] wsgi.secure_proxy_ssl_header = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.724325] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] wsgi.ssl_ca_file = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.724483] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] wsgi.ssl_cert_file = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.724642] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] wsgi.ssl_key_file = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.724806] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] wsgi.tcp_keepidle = 600 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.724983] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] wsgi.wsgi_log_format = %(client_ip)s "%(request_line)s" status: %(status_code)s len: %(body_length)s time: %(wall_seconds).7f {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.725164] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] zvm.ca_file = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.725328] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] zvm.cloud_connector_url = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.725608] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] zvm.image_tmp_path = /opt/stack/data/n-cpu-1/images {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.725783] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] zvm.reachable_timeout = 300 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.725963] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_policy.enforce_new_defaults = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.726368] env[62753]: WARNING oslo_config.cfg [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] Deprecated: Option "enforce_scope" from group "oslo_policy" is deprecated for removal (This configuration was added temporarily to facilitate a smooth transition to the new RBAC. OpenStack will always enforce scope checks. This configuration option is deprecated and will be removed in the 2025.2 cycle.). Its value may be silently ignored in the future. [ 525.726555] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_policy.enforce_scope = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.726732] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_policy.policy_default_rule = default {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.726914] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_policy.policy_dirs = ['policy.d'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.727100] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_policy.policy_file = policy.yaml {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.727279] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_policy.remote_content_type = application/x-www-form-urlencoded {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.727444] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_policy.remote_ssl_ca_crt_file = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.727605] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_policy.remote_ssl_client_crt_file = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.727763] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_policy.remote_ssl_client_key_file = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.727925] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_policy.remote_ssl_verify_server_crt = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.728104] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_versionedobjects.fatal_exception_format_errors = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.728280] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_middleware.http_basic_auth_user_file = /etc/htpasswd {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.728459] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] profiler.connection_string = messaging:// {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.728625] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] profiler.enabled = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.728794] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] profiler.es_doc_type = notification {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.728959] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] profiler.es_scroll_size = 10000 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.729161] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] profiler.es_scroll_time = 2m {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.729340] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] profiler.filter_error_trace = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.729517] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] profiler.hmac_keys = **** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.729688] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] profiler.sentinel_service_name = mymaster {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.729856] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] profiler.socket_timeout = 0.1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.730031] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] profiler.trace_requests = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.730223] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] profiler.trace_sqlalchemy = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.730419] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] profiler_jaeger.process_tags = {} {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.730584] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] profiler_jaeger.service_name_prefix = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.730750] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] profiler_otlp.service_name_prefix = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.730979] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] remote_debug.host = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.731176] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] remote_debug.port = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.731367] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.amqp_auto_delete = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.731534] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.amqp_durable_queues = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.731700] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.conn_pool_min_size = 2 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.731862] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.conn_pool_ttl = 1200 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.732033] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.direct_mandatory_flag = True {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.732203] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.enable_cancel_on_failover = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.732399] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.heartbeat_in_pthread = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.732689] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.heartbeat_rate = 3 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.732860] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.heartbeat_timeout_threshold = 60 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.733084] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.hostname = devstack {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.733300] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.kombu_compression = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.733506] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.kombu_failover_strategy = round-robin {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.733718] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.kombu_missing_consumer_retry_timeout = 60 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.733921] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.kombu_reconnect_delay = 1.0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.734152] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.processname = nova-compute {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.734368] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.rabbit_ha_queues = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.734577] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.rabbit_interval_max = 30 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.734790] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.rabbit_login_method = AMQPLAIN {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.734985] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.rabbit_qos_prefetch_count = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.735221] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.rabbit_quorum_delivery_limit = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.735434] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_bytes = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.735685] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.rabbit_quorum_max_memory_length = 0 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.735913] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.rabbit_quorum_queue = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.736163] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.rabbit_retry_backoff = 2 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.736374] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.rabbit_retry_interval = 1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.736565] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.rabbit_stream_fanout = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.736769] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.rabbit_transient_queues_ttl = 1800 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.736981] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.rabbit_transient_quorum_queue = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.737208] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.rpc_conn_pool_size = 30 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.737453] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.ssl = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.737682] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.ssl_ca_file = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.737901] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.ssl_cert_file = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.738130] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.ssl_enforce_fips_mode = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.738352] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.ssl_key_file = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.738564] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.ssl_version = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.738860] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_rabbit.use_queue_manager = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.739114] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_notifications.driver = ['messagingv2'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.739357] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_notifications.retry = -1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.739593] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_notifications.topics = ['notifications'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.739841] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_messaging_notifications.transport_url = **** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.740081] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.auth_section = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.740297] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.auth_type = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.740505] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.cafile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.740696] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.certfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.740924] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.collect_timing = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.741142] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.connect_retries = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.741340] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.connect_retry_delay = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.741584] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.endpoint_id = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.741787] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.endpoint_override = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.741959] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.insecure = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.742176] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.keyfile = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.742384] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.max_version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.742570] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.min_version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.742790] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.region_name = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.742976] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.retriable_status_codes = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.743192] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.service_name = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.743402] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.service_type = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.743609] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.split_loggers = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.743812] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.status_code_retries = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.744010] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.status_code_retry_delay = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.744186] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.timeout = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.744351] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.valid_interfaces = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.744513] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_limit.version = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.744681] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_reports.file_event_handler = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.744849] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_reports.file_event_handler_interval = 1 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.745023] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] oslo_reports.log_dir = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.745198] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vif_plug_linux_bridge_privileged.capabilities = [12] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.745365] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vif_plug_linux_bridge_privileged.group = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.745544] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vif_plug_linux_bridge_privileged.helper_command = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.745733] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vif_plug_linux_bridge_privileged.logger_name = oslo_privsep.daemon {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.745901] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vif_plug_linux_bridge_privileged.thread_pool_size = 8 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.746073] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vif_plug_linux_bridge_privileged.user = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.746252] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vif_plug_ovs_privileged.capabilities = [12, 1] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.746415] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vif_plug_ovs_privileged.group = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.746575] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vif_plug_ovs_privileged.helper_command = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.746742] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vif_plug_ovs_privileged.logger_name = oslo_privsep.daemon {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.746906] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vif_plug_ovs_privileged.thread_pool_size = 8 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.747075] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] vif_plug_ovs_privileged.user = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.747252] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_vif_linux_bridge.flat_interface = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.747437] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_vif_linux_bridge.forward_bridge_interface = ['all'] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.747611] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_vif_linux_bridge.iptables_bottom_regex = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.747783] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_vif_linux_bridge.iptables_drop_action = DROP {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.747955] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_vif_linux_bridge.iptables_top_regex = {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.748137] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_vif_linux_bridge.network_device_mtu = 1500 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.748307] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_vif_linux_bridge.use_ipv6 = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.748469] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_vif_linux_bridge.vlan_interface = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.748647] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_vif_ovs.default_qos_type = linux-noop {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.748820] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_vif_ovs.isolate_vif = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.748988] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_vif_ovs.network_device_mtu = 1500 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.749191] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_vif_ovs.ovs_vsctl_timeout = 120 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.749370] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_vif_ovs.ovsdb_connection = tcp:127.0.0.1:6640 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.749557] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_vif_ovs.ovsdb_interface = native {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.749750] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] os_vif_ovs.per_port_bridge = False {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.749930] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] privsep_osbrick.capabilities = [21] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.750106] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] privsep_osbrick.group = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.750270] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] privsep_osbrick.helper_command = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.750436] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] privsep_osbrick.logger_name = os_brick.privileged {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.750602] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] privsep_osbrick.thread_pool_size = 8 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.750758] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] privsep_osbrick.user = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.750931] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] nova_sys_admin.capabilities = [0, 1, 2, 3, 12, 21] {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.751103] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] nova_sys_admin.group = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.751265] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] nova_sys_admin.helper_command = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.751433] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] nova_sys_admin.logger_name = oslo_privsep.daemon {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.751619] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] nova_sys_admin.thread_pool_size = 8 {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.751783] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] nova_sys_admin.user = None {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2826}} [ 525.751918] env[62753]: DEBUG oslo_service.service [None req-b0eadb67-3098-41bf-9886-e7058966c6e2 None None] ******************************************************************************** {{(pid=62753) log_opt_values /opt/stack/data/venv/lib/python3.10/site-packages/oslo_config/cfg.py:2830}} [ 525.752420] env[62753]: INFO nova.service [-] Starting compute node (version 0.0.1) [ 526.258026] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Getting list of instances from cluster (obj){ [ 526.258026] env[62753]: value = "domain-c8" [ 526.258026] env[62753]: _type = "ClusterComputeResource" [ 526.258026] env[62753]: } {{(pid=62753) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 526.258026] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a7b07da-515b-4f3b-a170-bb6f9be58783 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 526.265488] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Got total of 0 instances {{(pid=62753) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 526.266124] env[62753]: WARNING nova.virt.vmwareapi.driver [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] The vmwareapi driver is not tested by the OpenStack project nor does it have clear maintainer(s) and thus its quality can not be ensured. It should be considered experimental and may be removed in a future release. If you are using the driver in production please let us know via the openstack-discuss mailing list. [ 526.266582] env[62753]: INFO nova.virt.node [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Generated node identity 42c8603e-76bc-4e29-ad77-cac0094f7308 [ 526.266805] env[62753]: INFO nova.virt.node [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Wrote node identity 42c8603e-76bc-4e29-ad77-cac0094f7308 to /opt/stack/data/n-cpu-1/compute_id [ 526.769430] env[62753]: WARNING nova.compute.manager [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Compute nodes ['42c8603e-76bc-4e29-ad77-cac0094f7308'] for host cpu-1 were not found in the database. If this is the first time this service is starting on this host, then you can ignore this warning. [ 527.775428] env[62753]: INFO nova.compute.manager [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Looking for unclaimed instances stuck in BUILDING status for nodes managed by this host [ 528.780897] env[62753]: WARNING nova.compute.manager [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] No compute node record found for host cpu-1. If this is the first time this service is starting on this host, then you can ignore this warning.: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host cpu-1 could not be found. [ 528.781292] env[62753]: DEBUG oslo_concurrency.lockutils [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.781426] env[62753]: DEBUG oslo_concurrency.lockutils [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 528.781555] env[62753]: DEBUG oslo_concurrency.lockutils [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 528.781737] env[62753]: DEBUG nova.compute.resource_tracker [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62753) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 528.782690] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c532733d-fa8c-4077-b780-f39c4262cbfe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.791201] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dddad0c4-33e6-4c79-be9c-644a53a3237e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.804644] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47d85a8f-d3c5-4b28-9092-c6c8b39d32f3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.810586] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c1d18fd-9699-4a19-85df-9ae23b4e82bf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 528.839923] env[62753]: DEBUG nova.compute.resource_tracker [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181498MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=62753) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 528.840094] env[62753]: DEBUG oslo_concurrency.lockutils [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 528.840332] env[62753]: DEBUG oslo_concurrency.lockutils [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 529.343864] env[62753]: WARNING nova.compute.resource_tracker [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] No compute node record for cpu-1:42c8603e-76bc-4e29-ad77-cac0094f7308: nova.exception_Remote.ComputeHostNotFound_Remote: Compute host 42c8603e-76bc-4e29-ad77-cac0094f7308 could not be found. [ 529.847470] env[62753]: INFO nova.compute.resource_tracker [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Compute node record created for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 with uuid: 42c8603e-76bc-4e29-ad77-cac0094f7308 [ 531.355376] env[62753]: DEBUG nova.compute.resource_tracker [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 531.355743] env[62753]: DEBUG nova.compute.resource_tracker [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 531.508552] env[62753]: INFO nova.scheduler.client.report [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] [req-5a2c79bb-4c10-4776-8b70-ac43ddd40098] Created resource provider record via placement API for resource provider with UUID 42c8603e-76bc-4e29-ad77-cac0094f7308 and name domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28. [ 531.524219] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-289eea7e-7123-491c-884c-3dd91ee3d1eb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.531469] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83c89097-d0ce-49f2-8ebc-1ec1405339df {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.560679] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48b7b21f-ee4f-44d4-a481-c36778c4b5d6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.567603] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43d7a58b-0cd5-4bc2-af6a-da9c87fbc107 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 531.580229] env[62753]: DEBUG nova.compute.provider_tree [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Updating inventory in ProviderTree for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 532.144019] env[62753]: DEBUG nova.scheduler.client.report [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Updated inventory for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 with generation 0 in Placement from set_inventory_for_provider using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:971}} [ 532.144019] env[62753]: DEBUG nova.compute.provider_tree [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Updating resource provider 42c8603e-76bc-4e29-ad77-cac0094f7308 generation from 0 to 1 during operation: update_inventory {{(pid=62753) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 532.144019] env[62753]: DEBUG nova.compute.provider_tree [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Updating inventory in ProviderTree for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 532.240020] env[62753]: DEBUG nova.compute.provider_tree [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Updating resource provider 42c8603e-76bc-4e29-ad77-cac0094f7308 generation from 1 to 2 during operation: update_traits {{(pid=62753) _update_generation /opt/stack/nova/nova/compute/provider_tree.py:164}} [ 532.743034] env[62753]: DEBUG nova.compute.resource_tracker [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 532.743312] env[62753]: DEBUG oslo_concurrency.lockutils [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 3.903s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 532.743312] env[62753]: DEBUG nova.service [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Creating RPC server for service compute {{(pid=62753) start /opt/stack/nova/nova/service.py:186}} [ 532.757861] env[62753]: DEBUG nova.service [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] Join ServiceGroup membership for this service compute {{(pid=62753) start /opt/stack/nova/nova/service.py:203}} [ 532.758180] env[62753]: DEBUG nova.servicegroup.drivers.db [None req-025f44cf-df45-4949-9bdc-2671099c0045 None None] DB_Driver: join new ServiceGroup member cpu-1 to the compute group, service = {{(pid=62753) join /opt/stack/nova/nova/servicegroup/drivers/db.py:44}} [ 560.764621] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._sync_power_states {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 561.268173] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Getting list of instances from cluster (obj){ [ 561.268173] env[62753]: value = "domain-c8" [ 561.268173] env[62753]: _type = "ClusterComputeResource" [ 561.268173] env[62753]: } {{(pid=62753) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 561.270284] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-723b28db-de8a-437a-a729-6e2848afe2cd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.278903] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Got total of 0 instances {{(pid=62753) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 561.279231] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._cleanup_running_deleted_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 561.279335] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Getting list of instances from cluster (obj){ [ 561.279335] env[62753]: value = "domain-c8" [ 561.279335] env[62753]: _type = "ClusterComputeResource" [ 561.279335] env[62753]: } {{(pid=62753) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 561.280218] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-baa49f1e-3a28-4077-8953-50f6584c9f77 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 561.287306] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Got total of 0 instances {{(pid=62753) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 569.078281] env[62753]: DEBUG oslo_concurrency.lockutils [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Acquiring lock "0f3de491-4b28-4a51-bdaa-884e57d925f3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 569.080079] env[62753]: DEBUG oslo_concurrency.lockutils [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Lock "0f3de491-4b28-4a51-bdaa-884e57d925f3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 569.582683] env[62753]: DEBUG nova.compute.manager [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 570.107764] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Acquiring lock "29dfb4db-c9ae-4a5c-8574-b27b13cdc83c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.108033] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Lock "29dfb4db-c9ae-4a5c-8574-b27b13cdc83c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.150584] env[62753]: DEBUG oslo_concurrency.lockutils [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 570.150864] env[62753]: DEBUG oslo_concurrency.lockutils [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 570.152620] env[62753]: INFO nova.compute.claims [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 570.612112] env[62753]: DEBUG nova.compute.manager [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 571.140702] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.239619] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa73084a-efc9-411d-a6fc-3b50e58b6059 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.247928] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d121503-836e-4736-84b8-fcc7003d1b63 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.287723] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9cb67485-ebff-4c3a-9943-25a790a12702 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.294209] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39c24c37-8440-4222-8213-9e41645ce778 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 571.310048] env[62753]: DEBUG nova.compute.provider_tree [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 571.473663] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Acquiring lock "f201df5d-dd05-4090-b6f3-bf099d8f919d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 571.473885] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Lock "f201df5d-dd05-4090-b6f3-bf099d8f919d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 571.819796] env[62753]: DEBUG nova.scheduler.client.report [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 571.978344] env[62753]: DEBUG nova.compute.manager [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 572.326656] env[62753]: DEBUG oslo_concurrency.lockutils [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.175s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 572.327162] env[62753]: DEBUG nova.compute.manager [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 572.330934] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.190s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 572.332049] env[62753]: INFO nova.compute.claims [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 572.516074] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 572.834852] env[62753]: DEBUG nova.compute.utils [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 572.834852] env[62753]: DEBUG nova.compute.manager [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 572.840403] env[62753]: DEBUG nova.network.neutron [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 573.348260] env[62753]: DEBUG nova.compute.manager [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 573.451040] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8167e258-139c-4b8b-b40b-b0645e3d7dee {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.461213] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3af67163-5aa6-46ab-aabb-4083d42f5804 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.494830] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a260085-ea7e-43bc-8f42-37fc0e80783e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.504225] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a98c979a-8a06-460b-a4c6-0f9c4c914269 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 573.520302] env[62753]: DEBUG nova.compute.provider_tree [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 573.880468] env[62753]: DEBUG nova.policy [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca7ddbd90a454b5ab41ebae1a1f4f33f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53e75036fc0d4d689316e9e4fde34731', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 574.027245] env[62753]: DEBUG nova.scheduler.client.report [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 574.392831] env[62753]: DEBUG nova.compute.manager [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 574.424694] env[62753]: DEBUG nova.virt.hardware [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 574.425021] env[62753]: DEBUG nova.virt.hardware [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 574.425183] env[62753]: DEBUG nova.virt.hardware [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 574.425369] env[62753]: DEBUG nova.virt.hardware [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 574.425513] env[62753]: DEBUG nova.virt.hardware [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 574.425664] env[62753]: DEBUG nova.virt.hardware [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 574.425882] env[62753]: DEBUG nova.virt.hardware [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 574.426058] env[62753]: DEBUG nova.virt.hardware [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 574.426469] env[62753]: DEBUG nova.virt.hardware [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 574.426641] env[62753]: DEBUG nova.virt.hardware [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 574.426850] env[62753]: DEBUG nova.virt.hardware [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 574.429503] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b153ab3-36e9-403c-9d65-28413cb4d48b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.444734] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6c11d55-699a-4290-a1fa-f7225453f883 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.465425] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf973a5d-d1f4-4697-a446-d433516c1f12 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 574.530580] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.200s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 574.531791] env[62753]: DEBUG nova.compute.manager [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 574.535781] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.020s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 574.538614] env[62753]: INFO nova.compute.claims [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 575.046171] env[62753]: DEBUG nova.compute.utils [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 575.047297] env[62753]: DEBUG nova.compute.manager [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Not allocating networking since 'none' was specified. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 575.548849] env[62753]: DEBUG nova.compute.manager [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 575.670446] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7fcd7aa7-8be1-46e9-9138-2b8d74248153 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.682478] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f4101a1-36c5-445c-b969-b7254b34d051 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.730982] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1736a8f3-3d2f-4cd3-a998-ae53af1cf0bf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.738990] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e36d2ff0-23ed-4571-a851-e8acf58f22fb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 575.754394] env[62753]: DEBUG nova.compute.provider_tree [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 576.257854] env[62753]: DEBUG nova.scheduler.client.report [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 576.452804] env[62753]: DEBUG nova.network.neutron [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Successfully created port: 9135d79d-2d1c-4591-b2dc-10e9e5983da2 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 576.566067] env[62753]: DEBUG nova.compute.manager [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 576.608693] env[62753]: DEBUG nova.virt.hardware [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 576.609029] env[62753]: DEBUG nova.virt.hardware [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 576.609215] env[62753]: DEBUG nova.virt.hardware [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 576.609456] env[62753]: DEBUG nova.virt.hardware [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 576.609615] env[62753]: DEBUG nova.virt.hardware [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 576.609895] env[62753]: DEBUG nova.virt.hardware [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 576.610114] env[62753]: DEBUG nova.virt.hardware [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 576.610330] env[62753]: DEBUG nova.virt.hardware [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 576.610543] env[62753]: DEBUG nova.virt.hardware [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 576.610834] env[62753]: DEBUG nova.virt.hardware [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 576.611818] env[62753]: DEBUG nova.virt.hardware [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 576.615444] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ccfd0ad-fbc8-4508-80f6-9cc6e2eed13b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.628857] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0712e65-739e-4556-b16d-8a6f310847c7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.648611] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Instance VIF info [] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 576.659411] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Creating folder: OpenStack. Parent ref: group-v4. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 576.659845] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-28452f75-e777-4115-b2cc-034863ce5308 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.673339] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Created folder: OpenStack in parent group-v4. [ 576.673572] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Creating folder: Project (45c117977ff14b4582f0b0317aca88b8). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 576.673846] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d29f6b94-221a-4be7-8013-2f0283b74e8a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.686655] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Created folder: Project (45c117977ff14b4582f0b0317aca88b8) in parent group-v284541. [ 576.686655] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Creating folder: Instances. Parent ref: group-v284542. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 576.686655] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d813a5fc-894a-4c10-99e2-d54b7aa951fa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.699635] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Created folder: Instances in parent group-v284542. [ 576.699890] env[62753]: DEBUG oslo.service.loopingcall [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 576.700131] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 576.700385] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-82ec26ae-b0c3-4dda-82b5-2f8f285268c6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 576.724303] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 576.724303] env[62753]: value = "task-1332074" [ 576.724303] env[62753]: _type = "Task" [ 576.724303] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 576.736828] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332074, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 576.763674] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.228s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 576.765056] env[62753]: DEBUG nova.compute.manager [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 577.235647] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332074, 'name': CreateVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.271165] env[62753]: DEBUG nova.compute.utils [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 577.273427] env[62753]: DEBUG nova.compute.manager [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 577.274049] env[62753]: DEBUG nova.network.neutron [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 577.386349] env[62753]: DEBUG nova.policy [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cebdec5ba33d459ebd6d1201e01b2618', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f25e5e09362b433d8e5a0eb596d96711', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 577.741973] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332074, 'name': CreateVM_Task, 'duration_secs': 0.55153} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 577.741973] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 577.741973] env[62753]: DEBUG oslo_vmware.service [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d72a3a1e-e5f8-4e12-9194-62cf6c10b77d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.748806] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 577.749784] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 577.749784] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 577.750124] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49caddd7-058d-45dc-a37b-e164caddfba8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 577.755401] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Waiting for the task: (returnval){ [ 577.755401] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52917bcf-2860-4bb0-9d81-498d6ad5d86d" [ 577.755401] env[62753]: _type = "Task" [ 577.755401] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 577.764036] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52917bcf-2860-4bb0-9d81-498d6ad5d86d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 577.774204] env[62753]: DEBUG nova.compute.manager [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 578.270478] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 578.271028] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 578.271028] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 578.271175] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 578.271847] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 578.272122] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c7e306cc-186a-4064-90d0-3ba9e9371b19 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.292664] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 578.292985] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 578.298426] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e428162-7061-41f2-b6f2-94712fc96dfe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.306424] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-08961c89-9fd5-49c5-9950-90288da4a2fd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.310952] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Waiting for the task: (returnval){ [ 578.310952] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f1266b-8b72-2261-561c-c3d76be3d47e" [ 578.310952] env[62753]: _type = "Task" [ 578.310952] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 578.320280] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f1266b-8b72-2261-561c-c3d76be3d47e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 578.333776] env[62753]: DEBUG nova.network.neutron [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Successfully created port: 403c03a8-05b0-4d79-8c2d-1b8156ced673 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 578.788040] env[62753]: DEBUG nova.compute.manager [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 578.825799] env[62753]: DEBUG nova.virt.hardware [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 578.825799] env[62753]: DEBUG nova.virt.hardware [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 578.825967] env[62753]: DEBUG nova.virt.hardware [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 578.826115] env[62753]: DEBUG nova.virt.hardware [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 578.826553] env[62753]: DEBUG nova.virt.hardware [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 578.826633] env[62753]: DEBUG nova.virt.hardware [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 578.827010] env[62753]: DEBUG nova.virt.hardware [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 578.827179] env[62753]: DEBUG nova.virt.hardware [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 578.829864] env[62753]: DEBUG nova.virt.hardware [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 578.829864] env[62753]: DEBUG nova.virt.hardware [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 578.829864] env[62753]: DEBUG nova.virt.hardware [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 578.829864] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Preparing fetch location {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 578.829864] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Creating directory with path [datastore2] vmware_temp/73e72d6f-cf44-4267-8b65-056a12f552f6/a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 578.830237] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8fe6446-3ec3-4cd4-be2d-ac64a30f0e17 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.833901] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a164f707-5523-49d3-8315-58e5465f9eb2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.843996] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f562ec56-6619-4899-948a-56246152f08a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.871393] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Created directory with path [datastore2] vmware_temp/73e72d6f-cf44-4267-8b65-056a12f552f6/a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 578.871393] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Fetch image to [datastore2] vmware_temp/73e72d6f-cf44-4267-8b65-056a12f552f6/a7541f15-bad6-4ea8-95ce-3499a4f01dda/tmp-sparse.vmdk {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 578.871393] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Downloading image file data a7541f15-bad6-4ea8-95ce-3499a4f01dda to [datastore2] vmware_temp/73e72d6f-cf44-4267-8b65-056a12f552f6/a7541f15-bad6-4ea8-95ce-3499a4f01dda/tmp-sparse.vmdk on the data store datastore2 {{(pid=62753) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 578.871393] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-608bd96f-55dd-4cfa-acf6-4efc42ee330c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.880302] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef00f1a5-a1c7-4da7-8fe4-9015c0c5fec6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.893249] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aa0bc71-32aa-465b-8938-204cf048fede {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.931895] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-502ac532-2c73-46a7-b69e-fbfc508c5a32 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 578.938284] env[62753]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-c0bea51d-c92a-4dec-ae8c-3c03db992a6e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.034494] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Downloading image file data a7541f15-bad6-4ea8-95ce-3499a4f01dda to the data store datastore2 {{(pid=62753) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 579.121021] env[62753]: DEBUG oslo_vmware.rw_handles [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/73e72d6f-cf44-4267-8b65-056a12f552f6/a7541f15-bad6-4ea8-95ce-3499a4f01dda/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62753) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 579.709688] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Acquiring lock "9afc71f1-21f3-422b-941f-23e29e29aa71" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 579.710012] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Lock "9afc71f1-21f3-422b-941f-23e29e29aa71" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 579.778971] env[62753]: DEBUG oslo_vmware.rw_handles [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Completed reading data from the image iterator. {{(pid=62753) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 579.779892] env[62753]: DEBUG oslo_vmware.rw_handles [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Closing write handle for https://esx7c2n3.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/73e72d6f-cf44-4267-8b65-056a12f552f6/a7541f15-bad6-4ea8-95ce-3499a4f01dda/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore2. {{(pid=62753) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 579.919515] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Downloaded image file data a7541f15-bad6-4ea8-95ce-3499a4f01dda to vmware_temp/73e72d6f-cf44-4267-8b65-056a12f552f6/a7541f15-bad6-4ea8-95ce-3499a4f01dda/tmp-sparse.vmdk on the data store datastore2 {{(pid=62753) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 579.921372] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Caching image {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 579.921372] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Copying Virtual Disk [datastore2] vmware_temp/73e72d6f-cf44-4267-8b65-056a12f552f6/a7541f15-bad6-4ea8-95ce-3499a4f01dda/tmp-sparse.vmdk to [datastore2] vmware_temp/73e72d6f-cf44-4267-8b65-056a12f552f6/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 579.921658] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-c10c03e1-244f-43f7-9d85-748a611ede3f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 579.930020] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Waiting for the task: (returnval){ [ 579.930020] env[62753]: value = "task-1332075" [ 579.930020] env[62753]: _type = "Task" [ 579.930020] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 579.942554] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332075, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.214822] env[62753]: DEBUG nova.compute.manager [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 580.441623] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332075, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 580.620204] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Acquiring lock "609fa528-f73f-4dbc-83a4-d19bb3d7d452" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.620204] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Lock "609fa528-f73f-4dbc-83a4-d19bb3d7d452" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.748126] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 580.748607] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 580.749999] env[62753]: INFO nova.compute.claims [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 580.941970] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332075, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.669413} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 580.942635] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Copied Virtual Disk [datastore2] vmware_temp/73e72d6f-cf44-4267-8b65-056a12f552f6/a7541f15-bad6-4ea8-95ce-3499a4f01dda/tmp-sparse.vmdk to [datastore2] vmware_temp/73e72d6f-cf44-4267-8b65-056a12f552f6/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 580.942967] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Deleting the datastore file [datastore2] vmware_temp/73e72d6f-cf44-4267-8b65-056a12f552f6/a7541f15-bad6-4ea8-95ce-3499a4f01dda/tmp-sparse.vmdk {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 580.943532] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1fe48140-601b-4b13-9bbd-307188bbe876 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 580.950447] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Waiting for the task: (returnval){ [ 580.950447] env[62753]: value = "task-1332076" [ 580.950447] env[62753]: _type = "Task" [ 580.950447] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 580.959503] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332076, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.019940] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 581.019940] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 581.019940] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Starting heal instance info cache {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 581.019940] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Rebuilding the list of instances to heal {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 581.072198] env[62753]: ERROR nova.compute.manager [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9135d79d-2d1c-4591-b2dc-10e9e5983da2, please check neutron logs for more information. [ 581.072198] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 581.072198] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.072198] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 581.072198] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.072198] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 581.072198] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.072198] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 581.072198] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.072198] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 581.072198] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.072198] env[62753]: ERROR nova.compute.manager raise self.value [ 581.072198] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.072198] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 581.072198] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.072198] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 581.072661] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.072661] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 581.072661] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9135d79d-2d1c-4591-b2dc-10e9e5983da2, please check neutron logs for more information. [ 581.072661] env[62753]: ERROR nova.compute.manager [ 581.073511] env[62753]: Traceback (most recent call last): [ 581.073511] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 581.073511] env[62753]: listener.cb(fileno) [ 581.073511] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.073511] env[62753]: result = function(*args, **kwargs) [ 581.073511] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 581.073511] env[62753]: return func(*args, **kwargs) [ 581.073511] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 581.073511] env[62753]: raise e [ 581.073511] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.073511] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 581.073511] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.073511] env[62753]: created_port_ids = self._update_ports_for_instance( [ 581.075025] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.075025] env[62753]: with excutils.save_and_reraise_exception(): [ 581.075025] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.075025] env[62753]: self.force_reraise() [ 581.075025] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.075025] env[62753]: raise self.value [ 581.075025] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.075025] env[62753]: updated_port = self._update_port( [ 581.075025] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.075025] env[62753]: _ensure_no_port_binding_failure(port) [ 581.075025] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.075025] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 581.075025] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 9135d79d-2d1c-4591-b2dc-10e9e5983da2, please check neutron logs for more information. [ 581.075025] env[62753]: Removing descriptor: 14 [ 581.076272] env[62753]: ERROR nova.compute.manager [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9135d79d-2d1c-4591-b2dc-10e9e5983da2, please check neutron logs for more information. [ 581.076272] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Traceback (most recent call last): [ 581.076272] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 581.076272] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] yield resources [ 581.076272] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 581.076272] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] self.driver.spawn(context, instance, image_meta, [ 581.076272] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 581.076272] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 581.076272] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 581.076272] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] vm_ref = self.build_virtual_machine(instance, [ 581.076272] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 581.076721] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] vif_infos = vmwarevif.get_vif_info(self._session, [ 581.076721] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 581.076721] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] for vif in network_info: [ 581.076721] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 581.076721] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] return self._sync_wrapper(fn, *args, **kwargs) [ 581.076721] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 581.076721] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] self.wait() [ 581.076721] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 581.076721] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] self[:] = self._gt.wait() [ 581.076721] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 581.076721] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] return self._exit_event.wait() [ 581.076721] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 581.076721] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] result = hub.switch() [ 581.077105] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 581.077105] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] return self.greenlet.switch() [ 581.077105] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 581.077105] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] result = function(*args, **kwargs) [ 581.077105] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 581.077105] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] return func(*args, **kwargs) [ 581.077105] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 581.077105] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] raise e [ 581.077105] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 581.077105] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] nwinfo = self.network_api.allocate_for_instance( [ 581.077105] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 581.077105] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] created_port_ids = self._update_ports_for_instance( [ 581.077105] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 581.079339] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] with excutils.save_and_reraise_exception(): [ 581.079339] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 581.079339] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] self.force_reraise() [ 581.079339] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 581.079339] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] raise self.value [ 581.079339] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 581.079339] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] updated_port = self._update_port( [ 581.079339] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 581.079339] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] _ensure_no_port_binding_failure(port) [ 581.079339] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 581.079339] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] raise exception.PortBindingFailed(port_id=port['id']) [ 581.079339] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] nova.exception.PortBindingFailed: Binding failed for port 9135d79d-2d1c-4591-b2dc-10e9e5983da2, please check neutron logs for more information. [ 581.079339] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] [ 581.079720] env[62753]: INFO nova.compute.manager [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Terminating instance [ 581.082348] env[62753]: DEBUG oslo_concurrency.lockutils [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Acquiring lock "refresh_cache-0f3de491-4b28-4a51-bdaa-884e57d925f3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 581.082348] env[62753]: DEBUG oslo_concurrency.lockutils [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Acquired lock "refresh_cache-0f3de491-4b28-4a51-bdaa-884e57d925f3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 581.082348] env[62753]: DEBUG nova.network.neutron [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 581.122407] env[62753]: DEBUG nova.compute.manager [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 581.464408] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332076, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023496} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 581.464551] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 581.464781] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Moving file from [datastore2] vmware_temp/73e72d6f-cf44-4267-8b65-056a12f552f6/a7541f15-bad6-4ea8-95ce-3499a4f01dda to [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda. {{(pid=62753) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 581.465037] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-454fcfa4-1e42-44e3-9d07-e5d4671e3855 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.472752] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Waiting for the task: (returnval){ [ 581.472752] env[62753]: value = "task-1332077" [ 581.472752] env[62753]: _type = "Task" [ 581.472752] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 581.482676] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332077, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 581.522814] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 581.523039] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 581.523181] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 581.523314] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 581.523435] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Didn't find any instances for network info cache update. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 581.523636] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 581.523900] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 581.524112] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 581.524303] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 581.524484] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 581.525494] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 581.525772] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62753) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 581.526331] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 581.637184] env[62753]: DEBUG nova.network.neutron [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 581.659548] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 581.868485] env[62753]: DEBUG nova.network.neutron [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 581.916104] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e22c431e-c544-41ea-8bdb-a223c6185a54 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.934055] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e30c5a79-a47d-450d-9bd4-7a833c412a61 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 581.989540] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb03c837-c828-41ce-8ff5-60e3c27a82ec {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.001027] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332077, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.023694} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.001354] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] File moved {{(pid=62753) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 582.001536] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Cleaning up location [datastore2] vmware_temp/73e72d6f-cf44-4267-8b65-056a12f552f6 {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 582.001696] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Deleting the datastore file [datastore2] vmware_temp/73e72d6f-cf44-4267-8b65-056a12f552f6 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 582.003548] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc53745-b04c-416f-b089-c8f5762d15e6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.007569] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-506d328b-d616-46f0-896f-064b56323a02 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.032143] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 582.032677] env[62753]: DEBUG nova.compute.provider_tree [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 582.039914] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Waiting for the task: (returnval){ [ 582.039914] env[62753]: value = "task-1332078" [ 582.039914] env[62753]: _type = "Task" [ 582.039914] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.049324] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332078, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.023573} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.049570] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 582.052214] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f8be61c6-30af-4ec1-af36-8426668f6ec6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.057090] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Waiting for the task: (returnval){ [ 582.057090] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52cd934d-22c0-14f4-2e41-af78bcb0351b" [ 582.057090] env[62753]: _type = "Task" [ 582.057090] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.066288] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52cd934d-22c0-14f4-2e41-af78bcb0351b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.374065] env[62753]: DEBUG oslo_concurrency.lockutils [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Releasing lock "refresh_cache-0f3de491-4b28-4a51-bdaa-884e57d925f3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.374516] env[62753]: DEBUG nova.compute.manager [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 582.375518] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 582.375518] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9b6f713f-2d88-446e-ac4e-c98b1402e28b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.386964] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17eb7fd7-d5b8-412b-b8e3-5c2c76f4057a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.418308] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 0f3de491-4b28-4a51-bdaa-884e57d925f3 could not be found. [ 582.418617] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 582.419064] env[62753]: INFO nova.compute.manager [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 582.419317] env[62753]: DEBUG oslo.service.loopingcall [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 582.419535] env[62753]: DEBUG nova.compute.manager [-] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 582.419627] env[62753]: DEBUG nova.network.neutron [-] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 582.439757] env[62753]: DEBUG nova.network.neutron [-] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 582.444098] env[62753]: ERROR nova.compute.manager [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 403c03a8-05b0-4d79-8c2d-1b8156ced673, please check neutron logs for more information. [ 582.444098] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 582.444098] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.444098] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 582.444098] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.444098] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 582.444098] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.444098] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 582.444098] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.444098] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 582.444098] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.444098] env[62753]: ERROR nova.compute.manager raise self.value [ 582.444098] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.444098] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 582.444098] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.444098] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 582.445191] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.445191] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 582.445191] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 403c03a8-05b0-4d79-8c2d-1b8156ced673, please check neutron logs for more information. [ 582.445191] env[62753]: ERROR nova.compute.manager [ 582.445191] env[62753]: Traceback (most recent call last): [ 582.445191] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 582.445191] env[62753]: listener.cb(fileno) [ 582.445191] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.445191] env[62753]: result = function(*args, **kwargs) [ 582.445191] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 582.445191] env[62753]: return func(*args, **kwargs) [ 582.445191] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.445191] env[62753]: raise e [ 582.445191] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.445191] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 582.445191] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.445191] env[62753]: created_port_ids = self._update_ports_for_instance( [ 582.445191] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.445191] env[62753]: with excutils.save_and_reraise_exception(): [ 582.445191] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.445191] env[62753]: self.force_reraise() [ 582.445191] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.445191] env[62753]: raise self.value [ 582.445191] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.445191] env[62753]: updated_port = self._update_port( [ 582.445191] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.445191] env[62753]: _ensure_no_port_binding_failure(port) [ 582.445191] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.445191] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 582.445933] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 403c03a8-05b0-4d79-8c2d-1b8156ced673, please check neutron logs for more information. [ 582.445933] env[62753]: Removing descriptor: 16 [ 582.445933] env[62753]: ERROR nova.compute.manager [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 403c03a8-05b0-4d79-8c2d-1b8156ced673, please check neutron logs for more information. [ 582.445933] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Traceback (most recent call last): [ 582.445933] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 582.445933] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] yield resources [ 582.445933] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 582.445933] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] self.driver.spawn(context, instance, image_meta, [ 582.445933] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 582.445933] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 582.445933] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 582.445933] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] vm_ref = self.build_virtual_machine(instance, [ 582.446265] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 582.446265] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] vif_infos = vmwarevif.get_vif_info(self._session, [ 582.446265] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 582.446265] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] for vif in network_info: [ 582.446265] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 582.446265] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] return self._sync_wrapper(fn, *args, **kwargs) [ 582.446265] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 582.446265] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] self.wait() [ 582.446265] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 582.446265] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] self[:] = self._gt.wait() [ 582.446265] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 582.446265] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] return self._exit_event.wait() [ 582.446265] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 582.446603] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] result = hub.switch() [ 582.446603] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 582.446603] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] return self.greenlet.switch() [ 582.446603] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 582.446603] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] result = function(*args, **kwargs) [ 582.446603] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 582.446603] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] return func(*args, **kwargs) [ 582.446603] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 582.446603] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] raise e [ 582.446603] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 582.446603] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] nwinfo = self.network_api.allocate_for_instance( [ 582.446603] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 582.446603] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] created_port_ids = self._update_ports_for_instance( [ 582.446993] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 582.446993] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] with excutils.save_and_reraise_exception(): [ 582.446993] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 582.446993] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] self.force_reraise() [ 582.446993] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 582.446993] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] raise self.value [ 582.446993] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 582.446993] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] updated_port = self._update_port( [ 582.446993] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 582.446993] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] _ensure_no_port_binding_failure(port) [ 582.446993] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 582.446993] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] raise exception.PortBindingFailed(port_id=port['id']) [ 582.447307] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] nova.exception.PortBindingFailed: Binding failed for port 403c03a8-05b0-4d79-8c2d-1b8156ced673, please check neutron logs for more information. [ 582.447307] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] [ 582.447307] env[62753]: INFO nova.compute.manager [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Terminating instance [ 582.449778] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Acquiring lock "refresh_cache-f201df5d-dd05-4090-b6f3-bf099d8f919d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 582.449778] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Acquired lock "refresh_cache-f201df5d-dd05-4090-b6f3-bf099d8f919d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 582.449778] env[62753]: DEBUG nova.network.neutron [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 582.542591] env[62753]: DEBUG nova.scheduler.client.report [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 582.574775] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52cd934d-22c0-14f4-2e41-af78bcb0351b, 'name': SearchDatastore_Task, 'duration_secs': 0.008472} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 582.575053] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 582.575297] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c/29dfb4db-c9ae-4a5c-8574-b27b13cdc83c.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 582.575547] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b66a9792-f9d6-495d-a880-5f02f9c5741c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 582.586765] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Waiting for the task: (returnval){ [ 582.586765] env[62753]: value = "task-1332079" [ 582.586765] env[62753]: _type = "Task" [ 582.586765] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 582.600183] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332079, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 582.947532] env[62753]: DEBUG nova.network.neutron [-] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.020426] env[62753]: DEBUG nova.compute.manager [req-2c2968e9-aeb0-40bf-aaf4-b594e87288d9 req-8971a811-5b76-4939-bfe4-5bc0713e21e8 service nova] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Received event network-changed-9135d79d-2d1c-4591-b2dc-10e9e5983da2 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 583.020667] env[62753]: DEBUG nova.compute.manager [req-2c2968e9-aeb0-40bf-aaf4-b594e87288d9 req-8971a811-5b76-4939-bfe4-5bc0713e21e8 service nova] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Refreshing instance network info cache due to event network-changed-9135d79d-2d1c-4591-b2dc-10e9e5983da2. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 583.020868] env[62753]: DEBUG oslo_concurrency.lockutils [req-2c2968e9-aeb0-40bf-aaf4-b594e87288d9 req-8971a811-5b76-4939-bfe4-5bc0713e21e8 service nova] Acquiring lock "refresh_cache-0f3de491-4b28-4a51-bdaa-884e57d925f3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 583.020868] env[62753]: DEBUG oslo_concurrency.lockutils [req-2c2968e9-aeb0-40bf-aaf4-b594e87288d9 req-8971a811-5b76-4939-bfe4-5bc0713e21e8 service nova] Acquired lock "refresh_cache-0f3de491-4b28-4a51-bdaa-884e57d925f3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 583.026426] env[62753]: DEBUG nova.network.neutron [req-2c2968e9-aeb0-40bf-aaf4-b594e87288d9 req-8971a811-5b76-4939-bfe4-5bc0713e21e8 service nova] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Refreshing network info cache for port 9135d79d-2d1c-4591-b2dc-10e9e5983da2 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 583.026996] env[62753]: DEBUG nova.network.neutron [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.056772] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.308s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 583.057354] env[62753]: DEBUG nova.compute.manager [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 583.059984] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.401s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.061574] env[62753]: INFO nova.compute.claims [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 583.100465] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332079, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494062} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.100802] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c/29dfb4db-c9ae-4a5c-8574-b27b13cdc83c.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 583.101086] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 583.101369] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e858ddc4-082c-406b-8bb1-9c515e54bb33 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.108059] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Waiting for the task: (returnval){ [ 583.108059] env[62753]: value = "task-1332080" [ 583.108059] env[62753]: _type = "Task" [ 583.108059] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.117862] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332080, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.195896] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Acquiring lock "a1df2c79-3afe-4bd7-89dd-c792a36bdca2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.197480] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Lock "a1df2c79-3afe-4bd7-89dd-c792a36bdca2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 583.244806] env[62753]: DEBUG nova.network.neutron [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.452891] env[62753]: INFO nova.compute.manager [-] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Took 1.03 seconds to deallocate network for instance. [ 583.458947] env[62753]: DEBUG nova.compute.claims [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 583.459068] env[62753]: DEBUG oslo_concurrency.lockutils [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 583.571748] env[62753]: DEBUG nova.compute.utils [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 583.574709] env[62753]: DEBUG nova.compute.manager [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 583.574882] env[62753]: DEBUG nova.network.neutron [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 583.578122] env[62753]: DEBUG nova.network.neutron [req-2c2968e9-aeb0-40bf-aaf4-b594e87288d9 req-8971a811-5b76-4939-bfe4-5bc0713e21e8 service nova] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 583.628997] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332080, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059988} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 583.628997] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 583.631154] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9341399f-e833-4b12-b86c-4649294bfc9c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.659177] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Reconfiguring VM instance instance-00000002 to attach disk [datastore2] 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c/29dfb4db-c9ae-4a5c-8574-b27b13cdc83c.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 583.660137] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-424a2668-086d-4140-8cf5-d406885063a2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.681966] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Waiting for the task: (returnval){ [ 583.681966] env[62753]: value = "task-1332081" [ 583.681966] env[62753]: _type = "Task" [ 583.681966] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 583.693139] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332081, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 583.698149] env[62753]: DEBUG nova.compute.manager [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 583.741388] env[62753]: DEBUG nova.network.neutron [req-2c2968e9-aeb0-40bf-aaf4-b594e87288d9 req-8971a811-5b76-4939-bfe4-5bc0713e21e8 service nova] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 583.747123] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Releasing lock "refresh_cache-f201df5d-dd05-4090-b6f3-bf099d8f919d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 583.747474] env[62753]: DEBUG nova.compute.manager [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 583.748391] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 583.748391] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-494d04d9-2674-437d-a466-ae43b22d52a0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.755240] env[62753]: DEBUG nova.policy [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '24478710d8814e09991b3cd262880442', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'debeba1a43f94497b3bd29699cea1b3e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 583.768628] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a49ed1b9-d421-4bab-950f-5686048cd315 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 583.795963] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f201df5d-dd05-4090-b6f3-bf099d8f919d could not be found. [ 583.798304] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 583.798304] env[62753]: INFO nova.compute.manager [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Took 0.05 seconds to destroy the instance on the hypervisor. [ 583.798304] env[62753]: DEBUG oslo.service.loopingcall [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 583.798304] env[62753]: DEBUG nova.compute.manager [-] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 583.798447] env[62753]: DEBUG nova.network.neutron [-] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 583.955502] env[62753]: DEBUG nova.network.neutron [-] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.075489] env[62753]: DEBUG nova.compute.manager [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 584.197398] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332081, 'name': ReconfigVM_Task, 'duration_secs': 0.277162} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.197893] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Reconfigured VM instance instance-00000002 to attach disk [datastore2] 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c/29dfb4db-c9ae-4a5c-8574-b27b13cdc83c.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 584.198754] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-787a57b1-0b80-4f25-9618-a3cfeb30f29f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.205089] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6224c42-8be6-48ce-a00e-76185b1fb509 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.219259] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a60d633-067f-4a80-bbd8-fbbae6167417 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.230343] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Waiting for the task: (returnval){ [ 584.230343] env[62753]: value = "task-1332082" [ 584.230343] env[62753]: _type = "Task" [ 584.230343] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.231415] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 584.269044] env[62753]: DEBUG oslo_concurrency.lockutils [req-2c2968e9-aeb0-40bf-aaf4-b594e87288d9 req-8971a811-5b76-4939-bfe4-5bc0713e21e8 service nova] Releasing lock "refresh_cache-0f3de491-4b28-4a51-bdaa-884e57d925f3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 584.271156] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12e94f94-509c-4d1e-85f5-33f737b0aa62 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.279669] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332082, 'name': Rename_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.285323] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f73208b8-2ba9-49b2-ab97-18d33b4e56da {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.302105] env[62753]: DEBUG nova.compute.provider_tree [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 584.380239] env[62753]: DEBUG nova.compute.manager [req-300ecfc6-f937-42cb-8c21-c1f23aff6567 req-76335d38-86dc-46b2-915f-f862b3d5662e service nova] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Received event network-changed-403c03a8-05b0-4d79-8c2d-1b8156ced673 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 584.380239] env[62753]: DEBUG nova.compute.manager [req-300ecfc6-f937-42cb-8c21-c1f23aff6567 req-76335d38-86dc-46b2-915f-f862b3d5662e service nova] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Refreshing instance network info cache due to event network-changed-403c03a8-05b0-4d79-8c2d-1b8156ced673. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 584.380239] env[62753]: DEBUG oslo_concurrency.lockutils [req-300ecfc6-f937-42cb-8c21-c1f23aff6567 req-76335d38-86dc-46b2-915f-f862b3d5662e service nova] Acquiring lock "refresh_cache-f201df5d-dd05-4090-b6f3-bf099d8f919d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 584.380239] env[62753]: DEBUG oslo_concurrency.lockutils [req-300ecfc6-f937-42cb-8c21-c1f23aff6567 req-76335d38-86dc-46b2-915f-f862b3d5662e service nova] Acquired lock "refresh_cache-f201df5d-dd05-4090-b6f3-bf099d8f919d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 584.380239] env[62753]: DEBUG nova.network.neutron [req-300ecfc6-f937-42cb-8c21-c1f23aff6567 req-76335d38-86dc-46b2-915f-f862b3d5662e service nova] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Refreshing network info cache for port 403c03a8-05b0-4d79-8c2d-1b8156ced673 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 584.462383] env[62753]: DEBUG nova.network.neutron [-] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 584.526024] env[62753]: DEBUG nova.network.neutron [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Successfully created port: 1c21050c-9372-483e-9755-65a89cebd58f {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 584.741737] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332082, 'name': Rename_Task, 'duration_secs': 0.133487} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 584.741997] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 584.742249] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-2629ba9d-6ae8-4c05-8300-8d31619ad7de {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 584.749145] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Waiting for the task: (returnval){ [ 584.749145] env[62753]: value = "task-1332083" [ 584.749145] env[62753]: _type = "Task" [ 584.749145] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 584.757942] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332083, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 584.806407] env[62753]: DEBUG nova.scheduler.client.report [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 584.920324] env[62753]: DEBUG nova.network.neutron [req-300ecfc6-f937-42cb-8c21-c1f23aff6567 req-76335d38-86dc-46b2-915f-f862b3d5662e service nova] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 584.968024] env[62753]: INFO nova.compute.manager [-] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Took 1.17 seconds to deallocate network for instance. [ 584.970520] env[62753]: DEBUG nova.compute.claims [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 584.970520] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.012686] env[62753]: DEBUG nova.network.neutron [req-300ecfc6-f937-42cb-8c21-c1f23aff6567 req-76335d38-86dc-46b2-915f-f862b3d5662e service nova] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 585.057429] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Acquiring lock "579053e4-d16f-42dc-99c6-69c2c0d68040" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.057429] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Lock "579053e4-d16f-42dc-99c6-69c2c0d68040" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.089152] env[62753]: DEBUG nova.compute.manager [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 585.112533] env[62753]: DEBUG nova.virt.hardware [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 585.112669] env[62753]: DEBUG nova.virt.hardware [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 585.114646] env[62753]: DEBUG nova.virt.hardware [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 585.114646] env[62753]: DEBUG nova.virt.hardware [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 585.114646] env[62753]: DEBUG nova.virt.hardware [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 585.114646] env[62753]: DEBUG nova.virt.hardware [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 585.114646] env[62753]: DEBUG nova.virt.hardware [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 585.115253] env[62753]: DEBUG nova.virt.hardware [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 585.115253] env[62753]: DEBUG nova.virt.hardware [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 585.115253] env[62753]: DEBUG nova.virt.hardware [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 585.115253] env[62753]: DEBUG nova.virt.hardware [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 585.117312] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-859560cb-51ec-4ac4-882d-1c1456550634 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.126929] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19f17558-432e-49a0-86cd-190415b5a908 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.262662] env[62753]: DEBUG oslo_vmware.api [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332083, 'name': PowerOnVM_Task, 'duration_secs': 0.468389} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 585.262662] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 585.262884] env[62753]: INFO nova.compute.manager [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Took 8.70 seconds to spawn the instance on the hypervisor. [ 585.263165] env[62753]: DEBUG nova.compute.manager [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 585.264564] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5212dbed-93c4-4a48-a918-4fe866617608 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.314603] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.254s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.315674] env[62753]: DEBUG nova.compute.manager [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 585.318274] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 3.286s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.318274] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 585.318403] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62753) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 585.318654] env[62753]: DEBUG oslo_concurrency.lockutils [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 1.860s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 585.322299] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1300d05e-60e3-4aad-a571-27ab5c52d42a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.332409] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e136ffad-eb05-46a1-9410-5f2782c64391 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.350582] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3fb704-6ef6-4a23-a223-7653906f7e22 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.358342] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9e2a61e-4d71-43bd-b24e-6c1b4cc29ad9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 585.391460] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181488MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=62753) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 585.391680] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 585.514459] env[62753]: DEBUG oslo_concurrency.lockutils [req-300ecfc6-f937-42cb-8c21-c1f23aff6567 req-76335d38-86dc-46b2-915f-f862b3d5662e service nova] Releasing lock "refresh_cache-f201df5d-dd05-4090-b6f3-bf099d8f919d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 585.562251] env[62753]: DEBUG nova.compute.manager [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 585.783780] env[62753]: INFO nova.compute.manager [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Took 14.67 seconds to build instance. [ 585.826850] env[62753]: DEBUG nova.compute.utils [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 585.830398] env[62753]: DEBUG nova.compute.manager [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 585.830398] env[62753]: DEBUG nova.network.neutron [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 585.944830] env[62753]: DEBUG nova.policy [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1e1db672dee7429987a821d3e507787b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '14438fda0c434f35ba4ea405a18d6144', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 585.988316] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-394ee33d-45ae-488e-bacc-6b418369c985 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.001537] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff912b15-0b8a-442f-b332-04e67b509640 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.036838] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d036ac-1c49-4999-9f45-14c3889c8156 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.046105] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d06029a-eb77-43b4-bbed-917460667308 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 586.062597] env[62753]: DEBUG nova.compute.provider_tree [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 586.081572] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.233231] env[62753]: DEBUG oslo_concurrency.lockutils [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Acquiring lock "8d09404e-1359-42be-8795-226fcaadf0a3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 586.233642] env[62753]: DEBUG oslo_concurrency.lockutils [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Lock "8d09404e-1359-42be-8795-226fcaadf0a3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 586.286048] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ac60af5-c841-4e92-9108-165f68714839 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Lock "29dfb4db-c9ae-4a5c-8574-b27b13cdc83c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.178s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 586.312247] env[62753]: DEBUG nova.network.neutron [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Successfully created port: 4f5c359b-daa7-4424-a6e2-e32a57c957f2 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 586.337024] env[62753]: DEBUG nova.compute.manager [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 586.386591] env[62753]: ERROR nova.compute.manager [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 1c21050c-9372-483e-9755-65a89cebd58f, please check neutron logs for more information. [ 586.386591] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 586.386591] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.386591] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 586.386591] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.386591] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 586.386591] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.386591] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 586.386591] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.386591] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 586.386591] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.386591] env[62753]: ERROR nova.compute.manager raise self.value [ 586.386591] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.386591] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 586.386591] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.386591] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 586.387842] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.387842] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 586.387842] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 1c21050c-9372-483e-9755-65a89cebd58f, please check neutron logs for more information. [ 586.387842] env[62753]: ERROR nova.compute.manager [ 586.387842] env[62753]: Traceback (most recent call last): [ 586.387842] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 586.387842] env[62753]: listener.cb(fileno) [ 586.387842] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.387842] env[62753]: result = function(*args, **kwargs) [ 586.387842] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 586.387842] env[62753]: return func(*args, **kwargs) [ 586.387842] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.387842] env[62753]: raise e [ 586.387842] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.387842] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 586.387842] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.387842] env[62753]: created_port_ids = self._update_ports_for_instance( [ 586.387842] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.387842] env[62753]: with excutils.save_and_reraise_exception(): [ 586.387842] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.387842] env[62753]: self.force_reraise() [ 586.387842] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.387842] env[62753]: raise self.value [ 586.387842] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.387842] env[62753]: updated_port = self._update_port( [ 586.387842] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.387842] env[62753]: _ensure_no_port_binding_failure(port) [ 586.387842] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.387842] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 586.388537] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 1c21050c-9372-483e-9755-65a89cebd58f, please check neutron logs for more information. [ 586.388537] env[62753]: Removing descriptor: 14 [ 586.388537] env[62753]: ERROR nova.compute.manager [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 1c21050c-9372-483e-9755-65a89cebd58f, please check neutron logs for more information. [ 586.388537] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Traceback (most recent call last): [ 586.388537] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 586.388537] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] yield resources [ 586.388537] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 586.388537] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] self.driver.spawn(context, instance, image_meta, [ 586.388537] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 586.388537] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] self._vmops.spawn(context, instance, image_meta, injected_files, [ 586.388537] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 586.388537] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] vm_ref = self.build_virtual_machine(instance, [ 586.388835] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 586.388835] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] vif_infos = vmwarevif.get_vif_info(self._session, [ 586.388835] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 586.388835] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] for vif in network_info: [ 586.388835] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 586.388835] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] return self._sync_wrapper(fn, *args, **kwargs) [ 586.388835] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 586.388835] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] self.wait() [ 586.388835] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 586.388835] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] self[:] = self._gt.wait() [ 586.388835] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 586.388835] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] return self._exit_event.wait() [ 586.388835] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 586.389228] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] result = hub.switch() [ 586.389228] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 586.389228] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] return self.greenlet.switch() [ 586.389228] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 586.389228] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] result = function(*args, **kwargs) [ 586.389228] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 586.389228] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] return func(*args, **kwargs) [ 586.389228] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 586.389228] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] raise e [ 586.389228] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 586.389228] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] nwinfo = self.network_api.allocate_for_instance( [ 586.389228] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 586.389228] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] created_port_ids = self._update_ports_for_instance( [ 586.389634] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 586.389634] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] with excutils.save_and_reraise_exception(): [ 586.389634] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 586.389634] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] self.force_reraise() [ 586.389634] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 586.389634] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] raise self.value [ 586.389634] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 586.389634] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] updated_port = self._update_port( [ 586.389634] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 586.389634] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] _ensure_no_port_binding_failure(port) [ 586.389634] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 586.389634] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] raise exception.PortBindingFailed(port_id=port['id']) [ 586.389930] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] nova.exception.PortBindingFailed: Binding failed for port 1c21050c-9372-483e-9755-65a89cebd58f, please check neutron logs for more information. [ 586.389930] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] [ 586.389930] env[62753]: INFO nova.compute.manager [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Terminating instance [ 586.394355] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Acquiring lock "refresh_cache-9afc71f1-21f3-422b-941f-23e29e29aa71" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 586.394510] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Acquired lock "refresh_cache-9afc71f1-21f3-422b-941f-23e29e29aa71" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 586.394828] env[62753]: DEBUG nova.network.neutron [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 586.567189] env[62753]: DEBUG nova.scheduler.client.report [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 586.736898] env[62753]: DEBUG nova.compute.manager [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 586.922883] env[62753]: DEBUG nova.network.neutron [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.075532] env[62753]: DEBUG oslo_concurrency.lockutils [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.756s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 587.075954] env[62753]: ERROR nova.compute.manager [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9135d79d-2d1c-4591-b2dc-10e9e5983da2, please check neutron logs for more information. [ 587.075954] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Traceback (most recent call last): [ 587.075954] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 587.075954] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] self.driver.spawn(context, instance, image_meta, [ 587.075954] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 587.075954] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 587.075954] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 587.075954] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] vm_ref = self.build_virtual_machine(instance, [ 587.075954] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 587.075954] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] vif_infos = vmwarevif.get_vif_info(self._session, [ 587.075954] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 587.076291] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] for vif in network_info: [ 587.076291] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 587.076291] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] return self._sync_wrapper(fn, *args, **kwargs) [ 587.076291] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 587.076291] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] self.wait() [ 587.076291] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 587.076291] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] self[:] = self._gt.wait() [ 587.076291] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 587.076291] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] return self._exit_event.wait() [ 587.076291] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 587.076291] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] result = hub.switch() [ 587.076291] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 587.076291] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] return self.greenlet.switch() [ 587.076601] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 587.076601] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] result = function(*args, **kwargs) [ 587.076601] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 587.076601] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] return func(*args, **kwargs) [ 587.076601] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 587.076601] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] raise e [ 587.076601] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 587.076601] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] nwinfo = self.network_api.allocate_for_instance( [ 587.076601] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 587.076601] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] created_port_ids = self._update_ports_for_instance( [ 587.076601] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 587.076601] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] with excutils.save_and_reraise_exception(): [ 587.076601] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 587.077226] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] self.force_reraise() [ 587.077226] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 587.077226] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] raise self.value [ 587.077226] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 587.077226] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] updated_port = self._update_port( [ 587.077226] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 587.077226] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] _ensure_no_port_binding_failure(port) [ 587.077226] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 587.077226] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] raise exception.PortBindingFailed(port_id=port['id']) [ 587.077226] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] nova.exception.PortBindingFailed: Binding failed for port 9135d79d-2d1c-4591-b2dc-10e9e5983da2, please check neutron logs for more information. [ 587.077226] env[62753]: ERROR nova.compute.manager [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] [ 587.077621] env[62753]: DEBUG nova.compute.utils [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Binding failed for port 9135d79d-2d1c-4591-b2dc-10e9e5983da2, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 587.077984] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.847s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 587.079808] env[62753]: INFO nova.compute.claims [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 587.084395] env[62753]: DEBUG nova.compute.manager [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Build of instance 0f3de491-4b28-4a51-bdaa-884e57d925f3 was re-scheduled: Binding failed for port 9135d79d-2d1c-4591-b2dc-10e9e5983da2, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 587.085047] env[62753]: DEBUG nova.compute.manager [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 587.085277] env[62753]: DEBUG oslo_concurrency.lockutils [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Acquiring lock "refresh_cache-0f3de491-4b28-4a51-bdaa-884e57d925f3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 587.085421] env[62753]: DEBUG oslo_concurrency.lockutils [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Acquired lock "refresh_cache-0f3de491-4b28-4a51-bdaa-884e57d925f3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 587.085577] env[62753]: DEBUG nova.network.neutron [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 587.114789] env[62753]: DEBUG nova.network.neutron [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.268691] env[62753]: DEBUG oslo_concurrency.lockutils [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 587.350248] env[62753]: DEBUG nova.compute.manager [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 587.382539] env[62753]: DEBUG nova.virt.hardware [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 587.382757] env[62753]: DEBUG nova.virt.hardware [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 587.382914] env[62753]: DEBUG nova.virt.hardware [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 587.383195] env[62753]: DEBUG nova.virt.hardware [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 587.383354] env[62753]: DEBUG nova.virt.hardware [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 587.383496] env[62753]: DEBUG nova.virt.hardware [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 587.383704] env[62753]: DEBUG nova.virt.hardware [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 587.383854] env[62753]: DEBUG nova.virt.hardware [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 587.384079] env[62753]: DEBUG nova.virt.hardware [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 587.384184] env[62753]: DEBUG nova.virt.hardware [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 587.384349] env[62753]: DEBUG nova.virt.hardware [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 587.385256] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2a03205-0358-4942-af79-c286988d8fb1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.395952] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae626b36-2cf7-4933-ac14-2019115cc733 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.620487] env[62753]: DEBUG nova.network.neutron [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.631403] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Releasing lock "refresh_cache-9afc71f1-21f3-422b-941f-23e29e29aa71" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 587.631786] env[62753]: DEBUG nova.compute.manager [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 587.632049] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 587.632897] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-fdb0d492-7031-443b-be98-adb70cef8a36 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.648772] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa552b64-56c2-4e40-b225-599b7c375770 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 587.676335] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 9afc71f1-21f3-422b-941f-23e29e29aa71 could not be found. [ 587.676635] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 587.676753] env[62753]: INFO nova.compute.manager [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Took 0.04 seconds to destroy the instance on the hypervisor. [ 587.677015] env[62753]: DEBUG oslo.service.loopingcall [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 587.677257] env[62753]: DEBUG nova.compute.manager [-] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 587.677428] env[62753]: DEBUG nova.network.neutron [-] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 587.729103] env[62753]: DEBUG nova.network.neutron [-] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 587.802453] env[62753]: DEBUG nova.network.neutron [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 587.992192] env[62753]: DEBUG nova.compute.manager [req-e6444a99-79dc-4cdf-b7a5-0d04ec00c3b8 req-7252dfbe-ec62-4ceb-9968-864bedfc37c7 service nova] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Received event network-vif-deleted-403c03a8-05b0-4d79-8c2d-1b8156ced673 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 588.072044] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Acquiring lock "4e1ae4a1-a06c-4afc-9c55-c9a728624ed9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.072044] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Lock "4e1ae4a1-a06c-4afc-9c55-c9a728624ed9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.232812] env[62753]: DEBUG nova.network.neutron [-] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.244370] env[62753]: ERROR nova.compute.manager [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4f5c359b-daa7-4424-a6e2-e32a57c957f2, please check neutron logs for more information. [ 588.244370] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 588.244370] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.244370] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 588.244370] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.244370] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 588.244370] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.244370] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 588.244370] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.244370] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 588.244370] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.244370] env[62753]: ERROR nova.compute.manager raise self.value [ 588.244370] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.244370] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 588.244370] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.244370] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 588.244929] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.244929] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 588.244929] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4f5c359b-daa7-4424-a6e2-e32a57c957f2, please check neutron logs for more information. [ 588.244929] env[62753]: ERROR nova.compute.manager [ 588.244929] env[62753]: Traceback (most recent call last): [ 588.244929] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 588.244929] env[62753]: listener.cb(fileno) [ 588.244929] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.244929] env[62753]: result = function(*args, **kwargs) [ 588.244929] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 588.244929] env[62753]: return func(*args, **kwargs) [ 588.244929] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.244929] env[62753]: raise e [ 588.244929] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.244929] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 588.244929] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.244929] env[62753]: created_port_ids = self._update_ports_for_instance( [ 588.244929] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.244929] env[62753]: with excutils.save_and_reraise_exception(): [ 588.244929] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.244929] env[62753]: self.force_reraise() [ 588.244929] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.244929] env[62753]: raise self.value [ 588.244929] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.244929] env[62753]: updated_port = self._update_port( [ 588.244929] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.244929] env[62753]: _ensure_no_port_binding_failure(port) [ 588.244929] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.244929] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 588.245707] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 4f5c359b-daa7-4424-a6e2-e32a57c957f2, please check neutron logs for more information. [ 588.245707] env[62753]: Removing descriptor: 16 [ 588.245707] env[62753]: ERROR nova.compute.manager [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4f5c359b-daa7-4424-a6e2-e32a57c957f2, please check neutron logs for more information. [ 588.245707] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Traceback (most recent call last): [ 588.245707] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 588.245707] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] yield resources [ 588.245707] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 588.245707] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] self.driver.spawn(context, instance, image_meta, [ 588.245707] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 588.245707] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] self._vmops.spawn(context, instance, image_meta, injected_files, [ 588.245707] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 588.245707] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] vm_ref = self.build_virtual_machine(instance, [ 588.246023] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 588.246023] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] vif_infos = vmwarevif.get_vif_info(self._session, [ 588.246023] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 588.246023] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] for vif in network_info: [ 588.246023] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 588.246023] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] return self._sync_wrapper(fn, *args, **kwargs) [ 588.246023] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 588.246023] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] self.wait() [ 588.246023] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 588.246023] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] self[:] = self._gt.wait() [ 588.246023] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 588.246023] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] return self._exit_event.wait() [ 588.246023] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 588.246417] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] result = hub.switch() [ 588.246417] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 588.246417] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] return self.greenlet.switch() [ 588.246417] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 588.246417] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] result = function(*args, **kwargs) [ 588.246417] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 588.246417] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] return func(*args, **kwargs) [ 588.246417] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 588.246417] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] raise e [ 588.246417] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 588.246417] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] nwinfo = self.network_api.allocate_for_instance( [ 588.246417] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 588.246417] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] created_port_ids = self._update_ports_for_instance( [ 588.246742] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 588.246742] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] with excutils.save_and_reraise_exception(): [ 588.246742] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 588.246742] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] self.force_reraise() [ 588.246742] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 588.246742] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] raise self.value [ 588.246742] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 588.246742] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] updated_port = self._update_port( [ 588.246742] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 588.246742] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] _ensure_no_port_binding_failure(port) [ 588.246742] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 588.246742] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] raise exception.PortBindingFailed(port_id=port['id']) [ 588.247039] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] nova.exception.PortBindingFailed: Binding failed for port 4f5c359b-daa7-4424-a6e2-e32a57c957f2, please check neutron logs for more information. [ 588.247039] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] [ 588.247039] env[62753]: INFO nova.compute.manager [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Terminating instance [ 588.248571] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Acquiring lock "refresh_cache-609fa528-f73f-4dbc-83a4-d19bb3d7d452" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 588.249267] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Acquired lock "refresh_cache-609fa528-f73f-4dbc-83a4-d19bb3d7d452" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 588.249267] env[62753]: DEBUG nova.network.neutron [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 588.284679] env[62753]: DEBUG nova.compute.manager [None req-fcec736a-96d3-468f-aaf2-725df04033fb tempest-ServerDiagnosticsV248Test-821750372 tempest-ServerDiagnosticsV248Test-821750372-project-admin] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 588.286636] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94426093-d998-407f-8e25-b0e4bd7a2f03 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.294935] env[62753]: INFO nova.compute.manager [None req-fcec736a-96d3-468f-aaf2-725df04033fb tempest-ServerDiagnosticsV248Test-821750372 tempest-ServerDiagnosticsV248Test-821750372-project-admin] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Retrieving diagnostics [ 588.298184] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1461c280-206c-4de8-a82b-981b59ba425a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.330767] env[62753]: DEBUG oslo_concurrency.lockutils [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Releasing lock "refresh_cache-0f3de491-4b28-4a51-bdaa-884e57d925f3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 588.331018] env[62753]: DEBUG nova.compute.manager [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 588.331208] env[62753]: DEBUG nova.compute.manager [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 588.331361] env[62753]: DEBUG nova.network.neutron [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 588.342348] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Acquiring lock "1ba742c4-9fa7-431e-83eb-abd86477ac24" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.342559] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Lock "1ba742c4-9fa7-431e-83eb-abd86477ac24" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 588.362308] env[62753]: DEBUG nova.network.neutron [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.379328] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b52302ff-a35d-4f27-aa4a-379ba30a7a0d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.388364] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68915704-8a18-45d6-a70b-c7ad50c5809d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.420944] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a3d4fec-3036-4331-800c-11a54250e8a2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.428460] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beb5250d-967c-4fd5-a2cf-e14a04ead9d1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 588.441897] env[62753]: DEBUG nova.compute.provider_tree [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 588.575229] env[62753]: DEBUG nova.compute.manager [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 588.738043] env[62753]: INFO nova.compute.manager [-] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Took 1.06 seconds to deallocate network for instance. [ 588.740585] env[62753]: DEBUG nova.compute.claims [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 588.740765] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 588.785830] env[62753]: DEBUG nova.network.neutron [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 588.844899] env[62753]: DEBUG nova.compute.manager [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 588.864815] env[62753]: DEBUG nova.network.neutron [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.891187] env[62753]: DEBUG nova.network.neutron [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 588.948376] env[62753]: DEBUG nova.scheduler.client.report [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 589.102575] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.368044] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 589.368578] env[62753]: INFO nova.compute.manager [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 0f3de491-4b28-4a51-bdaa-884e57d925f3] Took 1.04 seconds to deallocate network for instance. [ 589.395519] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Releasing lock "refresh_cache-609fa528-f73f-4dbc-83a4-d19bb3d7d452" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 589.399317] env[62753]: DEBUG nova.compute.manager [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 589.399317] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 589.399317] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e86ae852-c802-4fb8-a709-296502ff7d95 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.411859] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e0a645-a6e1-4ee8-91e7-85529d49edac {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 589.443681] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 609fa528-f73f-4dbc-83a4-d19bb3d7d452 could not be found. [ 589.443907] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 589.444101] env[62753]: INFO nova.compute.manager [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Took 0.05 seconds to destroy the instance on the hypervisor. [ 589.444378] env[62753]: DEBUG oslo.service.loopingcall [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 589.444650] env[62753]: DEBUG nova.compute.manager [-] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 589.444650] env[62753]: DEBUG nova.network.neutron [-] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 589.455634] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.377s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 589.455751] env[62753]: DEBUG nova.compute.manager [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 589.461144] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 4.491s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 589.552556] env[62753]: DEBUG nova.network.neutron [-] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 589.965823] env[62753]: DEBUG nova.compute.utils [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 589.965823] env[62753]: DEBUG nova.compute.manager [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 589.965823] env[62753]: DEBUG nova.network.neutron [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 590.054912] env[62753]: DEBUG nova.network.neutron [-] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 590.107773] env[62753]: DEBUG nova.policy [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '24478710d8814e09991b3cd262880442', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'debeba1a43f94497b3bd29699cea1b3e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 590.151912] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5c38475-2bb3-4c4f-bd02-b5e0e6d5e0f8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.160629] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e04fa0c7-ee42-48ce-a9c8-e650cde9d7b2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.201062] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0736d904-0c69-43bb-a052-59e4449ab422 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.209056] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946356ff-8f3a-483b-a935-27783287bd70 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 590.226831] env[62753]: DEBUG nova.compute.provider_tree [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 590.339099] env[62753]: DEBUG nova.compute.manager [req-37ff235c-f13a-419d-b9aa-e0005654f731 req-b45dca4f-3c52-4e69-b97f-fb8abbc1e0d0 service nova] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Received event network-changed-1c21050c-9372-483e-9755-65a89cebd58f {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 590.339222] env[62753]: DEBUG nova.compute.manager [req-37ff235c-f13a-419d-b9aa-e0005654f731 req-b45dca4f-3c52-4e69-b97f-fb8abbc1e0d0 service nova] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Refreshing instance network info cache due to event network-changed-1c21050c-9372-483e-9755-65a89cebd58f. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 590.339424] env[62753]: DEBUG oslo_concurrency.lockutils [req-37ff235c-f13a-419d-b9aa-e0005654f731 req-b45dca4f-3c52-4e69-b97f-fb8abbc1e0d0 service nova] Acquiring lock "refresh_cache-9afc71f1-21f3-422b-941f-23e29e29aa71" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 590.339565] env[62753]: DEBUG oslo_concurrency.lockutils [req-37ff235c-f13a-419d-b9aa-e0005654f731 req-b45dca4f-3c52-4e69-b97f-fb8abbc1e0d0 service nova] Acquired lock "refresh_cache-9afc71f1-21f3-422b-941f-23e29e29aa71" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 590.339785] env[62753]: DEBUG nova.network.neutron [req-37ff235c-f13a-419d-b9aa-e0005654f731 req-b45dca4f-3c52-4e69-b97f-fb8abbc1e0d0 service nova] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Refreshing network info cache for port 1c21050c-9372-483e-9755-65a89cebd58f {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 590.406647] env[62753]: INFO nova.scheduler.client.report [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Deleted allocations for instance 0f3de491-4b28-4a51-bdaa-884e57d925f3 [ 590.475074] env[62753]: DEBUG nova.compute.manager [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 590.561290] env[62753]: INFO nova.compute.manager [-] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Took 1.12 seconds to deallocate network for instance. [ 590.562899] env[62753]: DEBUG nova.compute.claims [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 590.563238] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.729714] env[62753]: DEBUG nova.scheduler.client.report [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 590.888775] env[62753]: DEBUG nova.network.neutron [req-37ff235c-f13a-419d-b9aa-e0005654f731 req-b45dca4f-3c52-4e69-b97f-fb8abbc1e0d0 service nova] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 590.891088] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Acquiring lock "3cd890d9-4f45-404d-ba3e-7104abcf2b7d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 590.891983] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Lock "3cd890d9-4f45-404d-ba3e-7104abcf2b7d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 590.914493] env[62753]: DEBUG oslo_concurrency.lockutils [None req-423f1ebd-1e18-48c7-95fb-cde809455a18 tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Lock "0f3de491-4b28-4a51-bdaa-884e57d925f3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.836s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.114239] env[62753]: DEBUG nova.network.neutron [req-37ff235c-f13a-419d-b9aa-e0005654f731 req-b45dca4f-3c52-4e69-b97f-fb8abbc1e0d0 service nova] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.162031] env[62753]: DEBUG nova.network.neutron [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Successfully created port: 10636605-321f-4ab2-9789-c183fe7592ff {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 591.178440] env[62753]: DEBUG nova.compute.manager [req-d87d6e2f-5c14-450d-bccb-f9086aa7bec8 req-f904cc9c-4650-45c1-8fde-16a4330d9592 service nova] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Received event network-changed-4f5c359b-daa7-4424-a6e2-e32a57c957f2 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 591.178440] env[62753]: DEBUG nova.compute.manager [req-d87d6e2f-5c14-450d-bccb-f9086aa7bec8 req-f904cc9c-4650-45c1-8fde-16a4330d9592 service nova] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Refreshing instance network info cache due to event network-changed-4f5c359b-daa7-4424-a6e2-e32a57c957f2. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 591.178635] env[62753]: DEBUG oslo_concurrency.lockutils [req-d87d6e2f-5c14-450d-bccb-f9086aa7bec8 req-f904cc9c-4650-45c1-8fde-16a4330d9592 service nova] Acquiring lock "refresh_cache-609fa528-f73f-4dbc-83a4-d19bb3d7d452" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.180082] env[62753]: DEBUG oslo_concurrency.lockutils [req-d87d6e2f-5c14-450d-bccb-f9086aa7bec8 req-f904cc9c-4650-45c1-8fde-16a4330d9592 service nova] Acquired lock "refresh_cache-609fa528-f73f-4dbc-83a4-d19bb3d7d452" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.180082] env[62753]: DEBUG nova.network.neutron [req-d87d6e2f-5c14-450d-bccb-f9086aa7bec8 req-f904cc9c-4650-45c1-8fde-16a4330d9592 service nova] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Refreshing network info cache for port 4f5c359b-daa7-4424-a6e2-e32a57c957f2 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 591.238248] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.777s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 591.238884] env[62753]: ERROR nova.compute.manager [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 403c03a8-05b0-4d79-8c2d-1b8156ced673, please check neutron logs for more information. [ 591.238884] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Traceback (most recent call last): [ 591.238884] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 591.238884] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] self.driver.spawn(context, instance, image_meta, [ 591.238884] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 591.238884] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 591.238884] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 591.238884] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] vm_ref = self.build_virtual_machine(instance, [ 591.238884] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 591.238884] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] vif_infos = vmwarevif.get_vif_info(self._session, [ 591.238884] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 591.239249] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] for vif in network_info: [ 591.239249] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 591.239249] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] return self._sync_wrapper(fn, *args, **kwargs) [ 591.239249] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 591.239249] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] self.wait() [ 591.239249] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 591.239249] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] self[:] = self._gt.wait() [ 591.239249] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 591.239249] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] return self._exit_event.wait() [ 591.239249] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 591.239249] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] result = hub.switch() [ 591.239249] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 591.239249] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] return self.greenlet.switch() [ 591.239624] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 591.239624] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] result = function(*args, **kwargs) [ 591.239624] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 591.239624] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] return func(*args, **kwargs) [ 591.239624] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 591.239624] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] raise e [ 591.239624] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 591.239624] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] nwinfo = self.network_api.allocate_for_instance( [ 591.239624] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 591.239624] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] created_port_ids = self._update_ports_for_instance( [ 591.239624] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 591.239624] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] with excutils.save_and_reraise_exception(): [ 591.239624] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 591.240201] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] self.force_reraise() [ 591.240201] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 591.240201] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] raise self.value [ 591.240201] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 591.240201] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] updated_port = self._update_port( [ 591.240201] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 591.240201] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] _ensure_no_port_binding_failure(port) [ 591.240201] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 591.240201] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] raise exception.PortBindingFailed(port_id=port['id']) [ 591.240201] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] nova.exception.PortBindingFailed: Binding failed for port 403c03a8-05b0-4d79-8c2d-1b8156ced673, please check neutron logs for more information. [ 591.240201] env[62753]: ERROR nova.compute.manager [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] [ 591.240534] env[62753]: DEBUG nova.compute.utils [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Binding failed for port 403c03a8-05b0-4d79-8c2d-1b8156ced673, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 591.240987] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 5.849s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.243343] env[62753]: DEBUG nova.compute.manager [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Build of instance f201df5d-dd05-4090-b6f3-bf099d8f919d was re-scheduled: Binding failed for port 403c03a8-05b0-4d79-8c2d-1b8156ced673, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 591.243932] env[62753]: DEBUG nova.compute.manager [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 591.244180] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Acquiring lock "refresh_cache-f201df5d-dd05-4090-b6f3-bf099d8f919d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 591.244328] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Acquired lock "refresh_cache-f201df5d-dd05-4090-b6f3-bf099d8f919d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 591.244489] env[62753]: DEBUG nova.network.neutron [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 591.397599] env[62753]: DEBUG nova.compute.manager [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 591.405316] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Acquiring lock "b4e2a5f2-7225-47fa-883b-d892c92e3949" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 591.405316] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Lock "b4e2a5f2-7225-47fa-883b-d892c92e3949" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 591.488672] env[62753]: DEBUG nova.compute.manager [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 591.545391] env[62753]: DEBUG nova.virt.hardware [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 591.545635] env[62753]: DEBUG nova.virt.hardware [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 591.545787] env[62753]: DEBUG nova.virt.hardware [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 591.545969] env[62753]: DEBUG nova.virt.hardware [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 591.546369] env[62753]: DEBUG nova.virt.hardware [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 591.546606] env[62753]: DEBUG nova.virt.hardware [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 591.546880] env[62753]: DEBUG nova.virt.hardware [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 591.547241] env[62753]: DEBUG nova.virt.hardware [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 591.547443] env[62753]: DEBUG nova.virt.hardware [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 591.547614] env[62753]: DEBUG nova.virt.hardware [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 591.547807] env[62753]: DEBUG nova.virt.hardware [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 591.549380] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0b8d088-99bc-4196-b63d-e88d8449c19c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.561457] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9303a96f-d66b-443d-924f-f4d27e900269 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 591.618859] env[62753]: DEBUG oslo_concurrency.lockutils [req-37ff235c-f13a-419d-b9aa-e0005654f731 req-b45dca4f-3c52-4e69-b97f-fb8abbc1e0d0 service nova] Releasing lock "refresh_cache-9afc71f1-21f3-422b-941f-23e29e29aa71" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 591.619056] env[62753]: DEBUG nova.compute.manager [req-37ff235c-f13a-419d-b9aa-e0005654f731 req-b45dca4f-3c52-4e69-b97f-fb8abbc1e0d0 service nova] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Received event network-vif-deleted-1c21050c-9372-483e-9755-65a89cebd58f {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 591.724215] env[62753]: DEBUG nova.network.neutron [req-d87d6e2f-5c14-450d-bccb-f9086aa7bec8 req-f904cc9c-4650-45c1-8fde-16a4330d9592 service nova] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.778657] env[62753]: DEBUG nova.network.neutron [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 591.911188] env[62753]: DEBUG nova.compute.manager [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 591.920812] env[62753]: DEBUG nova.network.neutron [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.928356] env[62753]: DEBUG nova.network.neutron [req-d87d6e2f-5c14-450d-bccb-f9086aa7bec8 req-f904cc9c-4650-45c1-8fde-16a4330d9592 service nova] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 591.944355] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.290047] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 592.431085] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Releasing lock "refresh_cache-f201df5d-dd05-4090-b6f3-bf099d8f919d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.434996] env[62753]: DEBUG nova.compute.manager [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 592.435170] env[62753]: DEBUG nova.compute.manager [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 592.435548] env[62753]: DEBUG nova.network.neutron [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 592.441712] env[62753]: DEBUG oslo_concurrency.lockutils [req-d87d6e2f-5c14-450d-bccb-f9086aa7bec8 req-f904cc9c-4650-45c1-8fde-16a4330d9592 service nova] Releasing lock "refresh_cache-609fa528-f73f-4dbc-83a4-d19bb3d7d452" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 592.441953] env[62753]: DEBUG nova.compute.manager [req-d87d6e2f-5c14-450d-bccb-f9086aa7bec8 req-f904cc9c-4650-45c1-8fde-16a4330d9592 service nova] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Received event network-vif-deleted-4f5c359b-daa7-4424-a6e2-e32a57c957f2 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 592.457449] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.482377] env[62753]: DEBUG nova.network.neutron [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 592.750380] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Acquiring lock "e5faf8e5-a5f6-4484-9304-7bbef03b33fc" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 592.750380] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Lock "e5faf8e5-a5f6-4484-9304-7bbef03b33fc" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 592.794790] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance f201df5d-dd05-4090-b6f3-bf099d8f919d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 592.795324] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 9afc71f1-21f3-422b-941f-23e29e29aa71 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 592.795809] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 609fa528-f73f-4dbc-83a4-d19bb3d7d452 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 592.797484] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance a1df2c79-3afe-4bd7-89dd-c792a36bdca2 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 592.985249] env[62753]: DEBUG nova.network.neutron [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 593.304496] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 579053e4-d16f-42dc-99c6-69c2c0d68040 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 593.491084] env[62753]: INFO nova.compute.manager [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] [instance: f201df5d-dd05-4090-b6f3-bf099d8f919d] Took 1.06 seconds to deallocate network for instance. [ 593.809491] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 8d09404e-1359-42be-8795-226fcaadf0a3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 594.318009] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 594.371996] env[62753]: ERROR nova.compute.manager [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 10636605-321f-4ab2-9789-c183fe7592ff, please check neutron logs for more information. [ 594.371996] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 594.371996] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.371996] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 594.371996] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.371996] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 594.371996] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.371996] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 594.371996] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.371996] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 594.371996] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.371996] env[62753]: ERROR nova.compute.manager raise self.value [ 594.371996] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.371996] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 594.371996] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.371996] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 594.372527] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.372527] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 594.372527] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 10636605-321f-4ab2-9789-c183fe7592ff, please check neutron logs for more information. [ 594.372527] env[62753]: ERROR nova.compute.manager [ 594.372527] env[62753]: Traceback (most recent call last): [ 594.372527] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 594.372527] env[62753]: listener.cb(fileno) [ 594.372527] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.372527] env[62753]: result = function(*args, **kwargs) [ 594.372527] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 594.372527] env[62753]: return func(*args, **kwargs) [ 594.372527] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.372527] env[62753]: raise e [ 594.372527] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.372527] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 594.372527] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.372527] env[62753]: created_port_ids = self._update_ports_for_instance( [ 594.372527] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.372527] env[62753]: with excutils.save_and_reraise_exception(): [ 594.372527] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.372527] env[62753]: self.force_reraise() [ 594.372527] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.372527] env[62753]: raise self.value [ 594.372527] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.372527] env[62753]: updated_port = self._update_port( [ 594.372527] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.372527] env[62753]: _ensure_no_port_binding_failure(port) [ 594.372527] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.372527] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 594.373377] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 10636605-321f-4ab2-9789-c183fe7592ff, please check neutron logs for more information. [ 594.373377] env[62753]: Removing descriptor: 16 [ 594.373377] env[62753]: ERROR nova.compute.manager [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 10636605-321f-4ab2-9789-c183fe7592ff, please check neutron logs for more information. [ 594.373377] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Traceback (most recent call last): [ 594.373377] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 594.373377] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] yield resources [ 594.373377] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 594.373377] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] self.driver.spawn(context, instance, image_meta, [ 594.373377] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 594.373377] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 594.373377] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 594.373377] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] vm_ref = self.build_virtual_machine(instance, [ 594.373831] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 594.373831] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] vif_infos = vmwarevif.get_vif_info(self._session, [ 594.373831] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 594.373831] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] for vif in network_info: [ 594.373831] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 594.373831] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] return self._sync_wrapper(fn, *args, **kwargs) [ 594.373831] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 594.373831] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] self.wait() [ 594.373831] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 594.373831] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] self[:] = self._gt.wait() [ 594.373831] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 594.373831] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] return self._exit_event.wait() [ 594.373831] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 594.374208] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] result = hub.switch() [ 594.374208] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 594.374208] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] return self.greenlet.switch() [ 594.374208] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 594.374208] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] result = function(*args, **kwargs) [ 594.374208] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 594.374208] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] return func(*args, **kwargs) [ 594.374208] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 594.374208] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] raise e [ 594.374208] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 594.374208] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] nwinfo = self.network_api.allocate_for_instance( [ 594.374208] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 594.374208] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] created_port_ids = self._update_ports_for_instance( [ 594.374719] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 594.374719] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] with excutils.save_and_reraise_exception(): [ 594.374719] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 594.374719] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] self.force_reraise() [ 594.374719] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 594.374719] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] raise self.value [ 594.374719] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 594.374719] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] updated_port = self._update_port( [ 594.374719] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 594.374719] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] _ensure_no_port_binding_failure(port) [ 594.374719] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 594.374719] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] raise exception.PortBindingFailed(port_id=port['id']) [ 594.375448] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] nova.exception.PortBindingFailed: Binding failed for port 10636605-321f-4ab2-9789-c183fe7592ff, please check neutron logs for more information. [ 594.375448] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] [ 594.375448] env[62753]: INFO nova.compute.manager [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Terminating instance [ 594.376066] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Acquiring lock "refresh_cache-a1df2c79-3afe-4bd7-89dd-c792a36bdca2" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.376483] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Acquired lock "refresh_cache-a1df2c79-3afe-4bd7-89dd-c792a36bdca2" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 594.376705] env[62753]: DEBUG nova.network.neutron [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 594.533459] env[62753]: INFO nova.scheduler.client.report [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Deleted allocations for instance f201df5d-dd05-4090-b6f3-bf099d8f919d [ 594.759170] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "11cf82b0-ea78-40b2-b737-dea8b8703ac3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.759453] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "11cf82b0-ea78-40b2-b737-dea8b8703ac3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.821564] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 1ba742c4-9fa7-431e-83eb-abd86477ac24 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 594.826564] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "32dae012-7743-4efb-b39b-820c9ed42cc5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 594.826785] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "32dae012-7743-4efb-b39b-820c9ed42cc5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 594.918677] env[62753]: DEBUG nova.compute.manager [req-35dc8f28-4b14-4d06-ad98-3b725810ec88 req-9b3bcdd9-d977-4ce6-b274-9184f88408c3 service nova] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Received event network-changed-10636605-321f-4ab2-9789-c183fe7592ff {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 594.918869] env[62753]: DEBUG nova.compute.manager [req-35dc8f28-4b14-4d06-ad98-3b725810ec88 req-9b3bcdd9-d977-4ce6-b274-9184f88408c3 service nova] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Refreshing instance network info cache due to event network-changed-10636605-321f-4ab2-9789-c183fe7592ff. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 594.919079] env[62753]: DEBUG oslo_concurrency.lockutils [req-35dc8f28-4b14-4d06-ad98-3b725810ec88 req-9b3bcdd9-d977-4ce6-b274-9184f88408c3 service nova] Acquiring lock "refresh_cache-a1df2c79-3afe-4bd7-89dd-c792a36bdca2" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 594.924737] env[62753]: DEBUG nova.network.neutron [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.044201] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7929877a-cdd0-43fd-a2dd-0a7ca941bf02 tempest-ServerDiagnosticsNegativeTest-1460141877 tempest-ServerDiagnosticsNegativeTest-1460141877-project-member] Lock "f201df5d-dd05-4090-b6f3-bf099d8f919d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.568s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 595.156983] env[62753]: DEBUG nova.network.neutron [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 595.327058] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 3cd890d9-4f45-404d-ba3e-7104abcf2b7d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 595.548116] env[62753]: DEBUG nova.compute.manager [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 595.659834] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Releasing lock "refresh_cache-a1df2c79-3afe-4bd7-89dd-c792a36bdca2" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 595.660318] env[62753]: DEBUG nova.compute.manager [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 595.660514] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 595.660904] env[62753]: DEBUG oslo_concurrency.lockutils [req-35dc8f28-4b14-4d06-ad98-3b725810ec88 req-9b3bcdd9-d977-4ce6-b274-9184f88408c3 service nova] Acquired lock "refresh_cache-a1df2c79-3afe-4bd7-89dd-c792a36bdca2" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 595.662583] env[62753]: DEBUG nova.network.neutron [req-35dc8f28-4b14-4d06-ad98-3b725810ec88 req-9b3bcdd9-d977-4ce6-b274-9184f88408c3 service nova] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Refreshing network info cache for port 10636605-321f-4ab2-9789-c183fe7592ff {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 595.662583] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6962a013-3ea4-44b3-8673-bfd3ff02438d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.682530] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec85b4d2-1fca-45ca-b582-3939f29c72a4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 595.711802] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a1df2c79-3afe-4bd7-89dd-c792a36bdca2 could not be found. [ 595.711802] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 595.711802] env[62753]: INFO nova.compute.manager [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Took 0.05 seconds to destroy the instance on the hypervisor. [ 595.714287] env[62753]: DEBUG oslo.service.loopingcall [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 595.714287] env[62753]: DEBUG nova.compute.manager [-] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 595.714287] env[62753]: DEBUG nova.network.neutron [-] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 595.750401] env[62753]: DEBUG nova.network.neutron [-] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 595.833313] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance b4e2a5f2-7225-47fa-883b-d892c92e3949 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 596.075113] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.230244] env[62753]: DEBUG nova.network.neutron [req-35dc8f28-4b14-4d06-ad98-3b725810ec88 req-9b3bcdd9-d977-4ce6-b274-9184f88408c3 service nova] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 596.241285] env[62753]: DEBUG oslo_concurrency.lockutils [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Acquiring lock "2d028151-3d84-4195-9f93-28287dbfda09" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.241508] env[62753]: DEBUG oslo_concurrency.lockutils [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Lock "2d028151-3d84-4195-9f93-28287dbfda09" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 596.255850] env[62753]: DEBUG nova.network.neutron [-] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.335493] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance e5faf8e5-a5f6-4484-9304-7bbef03b33fc has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 596.336858] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 596.339019] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1280MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 596.406648] env[62753]: DEBUG nova.network.neutron [req-35dc8f28-4b14-4d06-ad98-3b725810ec88 req-9b3bcdd9-d977-4ce6-b274-9184f88408c3 service nova] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 596.617254] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36820127-598d-44e2-b0d7-bcaf25fde4aa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.627735] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a148d29d-264c-4e53-884b-3eaafef40783 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.664443] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbe14d5f-5512-4605-bbf1-fcac94b2f71a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.672227] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e7077e1-7144-45e4-b3b9-b55ec552efd5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 596.686857] env[62753]: DEBUG nova.compute.provider_tree [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 596.766326] env[62753]: INFO nova.compute.manager [-] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Took 1.05 seconds to deallocate network for instance. [ 596.767029] env[62753]: DEBUG nova.compute.claims [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 596.767378] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 596.912105] env[62753]: DEBUG oslo_concurrency.lockutils [req-35dc8f28-4b14-4d06-ad98-3b725810ec88 req-9b3bcdd9-d977-4ce6-b274-9184f88408c3 service nova] Releasing lock "refresh_cache-a1df2c79-3afe-4bd7-89dd-c792a36bdca2" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 597.042833] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Acquiring lock "989db77e-169f-4b3b-a0b9-c4417f4e89c4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.043085] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Lock "989db77e-169f-4b3b-a0b9-c4417f4e89c4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.195918] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 597.199088] env[62753]: DEBUG oslo_concurrency.lockutils [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Acquiring lock "3c8e531e-033c-4857-9865-b1024e143d44" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.199393] env[62753]: DEBUG oslo_concurrency.lockutils [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Lock "3c8e531e-033c-4857-9865-b1024e143d44" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.702837] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 597.702837] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 6.462s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 597.704623] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 11.622s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 597.705916] env[62753]: INFO nova.compute.claims [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 597.980309] env[62753]: DEBUG oslo_concurrency.lockutils [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Acquiring lock "e730e4f0-fe02-4926-8263-93c8a3065962" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 597.980710] env[62753]: DEBUG oslo_concurrency.lockutils [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Lock "e730e4f0-fe02-4926-8263-93c8a3065962" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.502875] env[62753]: DEBUG nova.compute.manager [req-b31c8b3e-04df-4530-acac-afb7a9a9582c req-4f8116eb-7a38-4484-b634-6e15a3e0964d service nova] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Received event network-vif-deleted-10636605-321f-4ab2-9789-c183fe7592ff {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 598.655781] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Acquiring lock "e8f80847-a14d-4abd-af9f-84c62dd1e395" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.655781] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Lock "e8f80847-a14d-4abd-af9f-84c62dd1e395" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 598.969531] env[62753]: DEBUG oslo_concurrency.lockutils [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Acquiring lock "3d13009d-5e65-43e1-87b3-3cb3cfd014e6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 598.969598] env[62753]: DEBUG oslo_concurrency.lockutils [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Lock "3d13009d-5e65-43e1-87b3-3cb3cfd014e6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.039710] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cfd8985-1257-4d37-8ca7-c692a6df1138 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.048481] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c2459a-7091-4992-b4be-0a07cb1c0f2f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.080025] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb91138-40cc-458b-ae94-5a58e2261e6d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.090705] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a8a09e-97b3-4048-bee8-c4b0637edf7a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 599.108412] env[62753]: DEBUG nova.compute.provider_tree [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 599.606784] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Acquiring lock "b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 599.606784] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Lock "b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 599.611074] env[62753]: DEBUG nova.scheduler.client.report [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 600.023810] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "a408799d-29a9-442f-9769-8f8438af8ae0" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 600.024105] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "a408799d-29a9-442f-9769-8f8438af8ae0" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.118019] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.413s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 600.118019] env[62753]: DEBUG nova.compute.manager [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 600.125072] env[62753]: DEBUG oslo_concurrency.lockutils [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.856s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 600.128294] env[62753]: INFO nova.compute.claims [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 600.362469] env[62753]: DEBUG nova.compute.manager [None req-3aa937f9-cc4f-403a-82da-4a5336219ace tempest-ServerDiagnosticsV248Test-821750372 tempest-ServerDiagnosticsV248Test-821750372-project-admin] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 600.363600] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0dd6bd0-9e29-445b-9d28-c6e0feda1739 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.374037] env[62753]: INFO nova.compute.manager [None req-3aa937f9-cc4f-403a-82da-4a5336219ace tempest-ServerDiagnosticsV248Test-821750372 tempest-ServerDiagnosticsV248Test-821750372-project-admin] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Retrieving diagnostics [ 600.376213] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb7c7329-87fa-4275-ae6a-0cb8a1c1a304 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 600.633729] env[62753]: DEBUG nova.compute.utils [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 600.635669] env[62753]: DEBUG nova.compute.manager [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 600.635855] env[62753]: DEBUG nova.network.neutron [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 600.712424] env[62753]: DEBUG nova.policy [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cd92e8f419e648628958c9d0ecb2d651', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ee88c21c59ef4869afad7e123fb84d42', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 601.139503] env[62753]: DEBUG nova.network.neutron [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Successfully created port: 08b654e9-1488-4940-b8c8-5528f7f78709 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 601.140343] env[62753]: DEBUG nova.compute.manager [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 601.494488] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45df7648-362a-486c-9807-795ac5183859 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.505095] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-050d7677-75ce-4f0b-b27c-90f841bc6daa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.541358] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7257cb02-38fc-4ce9-9b18-8d18012a90bc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.550074] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6edde55a-54e5-435e-9121-7d15e5a92d99 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 601.563761] env[62753]: DEBUG nova.compute.provider_tree [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 602.074500] env[62753]: DEBUG nova.scheduler.client.report [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 602.101874] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Acquiring lock "29dfb4db-c9ae-4a5c-8574-b27b13cdc83c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.102307] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Lock "29dfb4db-c9ae-4a5c-8574-b27b13cdc83c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.102596] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Acquiring lock "29dfb4db-c9ae-4a5c-8574-b27b13cdc83c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 602.102842] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Lock "29dfb4db-c9ae-4a5c-8574-b27b13cdc83c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.103099] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Lock "29dfb4db-c9ae-4a5c-8574-b27b13cdc83c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.106023] env[62753]: INFO nova.compute.manager [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Terminating instance [ 602.107301] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Acquiring lock "refresh_cache-29dfb4db-c9ae-4a5c-8574-b27b13cdc83c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.107472] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Acquired lock "refresh_cache-29dfb4db-c9ae-4a5c-8574-b27b13cdc83c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.107646] env[62753]: DEBUG nova.network.neutron [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 602.156792] env[62753]: DEBUG nova.compute.manager [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 602.193642] env[62753]: DEBUG nova.virt.hardware [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 602.195099] env[62753]: DEBUG nova.virt.hardware [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 602.195099] env[62753]: DEBUG nova.virt.hardware [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 602.195099] env[62753]: DEBUG nova.virt.hardware [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 602.195260] env[62753]: DEBUG nova.virt.hardware [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 602.195513] env[62753]: DEBUG nova.virt.hardware [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 602.195840] env[62753]: DEBUG nova.virt.hardware [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 602.196131] env[62753]: DEBUG nova.virt.hardware [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 602.196398] env[62753]: DEBUG nova.virt.hardware [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 602.196672] env[62753]: DEBUG nova.virt.hardware [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 602.198444] env[62753]: DEBUG nova.virt.hardware [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 602.198444] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a073eb5-b045-4f30-833f-fbc9cd2a3eae {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.210198] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6116b5bd-d552-423b-a52c-bdc8fe1d81a8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 602.491579] env[62753]: ERROR nova.compute.manager [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 08b654e9-1488-4940-b8c8-5528f7f78709, please check neutron logs for more information. [ 602.491579] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 602.491579] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.491579] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 602.491579] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.491579] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 602.491579] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.491579] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 602.491579] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.491579] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 602.491579] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.491579] env[62753]: ERROR nova.compute.manager raise self.value [ 602.491579] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.491579] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 602.491579] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.491579] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 602.492069] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.492069] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 602.492069] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 08b654e9-1488-4940-b8c8-5528f7f78709, please check neutron logs for more information. [ 602.492069] env[62753]: ERROR nova.compute.manager [ 602.493107] env[62753]: Traceback (most recent call last): [ 602.493157] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 602.493157] env[62753]: listener.cb(fileno) [ 602.493157] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.493157] env[62753]: result = function(*args, **kwargs) [ 602.493157] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 602.493157] env[62753]: return func(*args, **kwargs) [ 602.493157] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.493157] env[62753]: raise e [ 602.493157] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.493157] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 602.493157] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.493157] env[62753]: created_port_ids = self._update_ports_for_instance( [ 602.493157] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.493157] env[62753]: with excutils.save_and_reraise_exception(): [ 602.493157] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.493157] env[62753]: self.force_reraise() [ 602.493157] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.493157] env[62753]: raise self.value [ 602.493157] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.493157] env[62753]: updated_port = self._update_port( [ 602.493157] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.493157] env[62753]: _ensure_no_port_binding_failure(port) [ 602.493157] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.493157] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 602.493827] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 08b654e9-1488-4940-b8c8-5528f7f78709, please check neutron logs for more information. [ 602.493827] env[62753]: Removing descriptor: 16 [ 602.494033] env[62753]: ERROR nova.compute.manager [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 08b654e9-1488-4940-b8c8-5528f7f78709, please check neutron logs for more information. [ 602.494033] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Traceback (most recent call last): [ 602.494033] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 602.494033] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] yield resources [ 602.494033] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 602.494033] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] self.driver.spawn(context, instance, image_meta, [ 602.494033] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 602.494033] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] self._vmops.spawn(context, instance, image_meta, injected_files, [ 602.494033] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 602.494033] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] vm_ref = self.build_virtual_machine(instance, [ 602.494033] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 602.494462] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] vif_infos = vmwarevif.get_vif_info(self._session, [ 602.494462] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 602.494462] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] for vif in network_info: [ 602.494462] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 602.494462] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] return self._sync_wrapper(fn, *args, **kwargs) [ 602.494462] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 602.494462] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] self.wait() [ 602.494462] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 602.494462] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] self[:] = self._gt.wait() [ 602.494462] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 602.494462] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] return self._exit_event.wait() [ 602.494462] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 602.494462] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] result = hub.switch() [ 602.494892] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 602.494892] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] return self.greenlet.switch() [ 602.494892] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 602.494892] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] result = function(*args, **kwargs) [ 602.494892] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 602.494892] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] return func(*args, **kwargs) [ 602.494892] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 602.494892] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] raise e [ 602.494892] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 602.494892] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] nwinfo = self.network_api.allocate_for_instance( [ 602.494892] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 602.494892] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] created_port_ids = self._update_ports_for_instance( [ 602.494892] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 602.495265] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] with excutils.save_and_reraise_exception(): [ 602.495265] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 602.495265] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] self.force_reraise() [ 602.495265] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 602.495265] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] raise self.value [ 602.495265] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 602.495265] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] updated_port = self._update_port( [ 602.495265] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 602.495265] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] _ensure_no_port_binding_failure(port) [ 602.495265] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 602.495265] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] raise exception.PortBindingFailed(port_id=port['id']) [ 602.495265] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] nova.exception.PortBindingFailed: Binding failed for port 08b654e9-1488-4940-b8c8-5528f7f78709, please check neutron logs for more information. [ 602.495265] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] [ 602.495713] env[62753]: INFO nova.compute.manager [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Terminating instance [ 602.503495] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Acquiring lock "refresh_cache-579053e4-d16f-42dc-99c6-69c2c0d68040" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 602.503672] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Acquired lock "refresh_cache-579053e4-d16f-42dc-99c6-69c2c0d68040" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 602.503842] env[62753]: DEBUG nova.network.neutron [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 602.580543] env[62753]: DEBUG oslo_concurrency.lockutils [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.454s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 602.580543] env[62753]: DEBUG nova.compute.manager [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 602.583256] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.842s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 602.653269] env[62753]: DEBUG nova.network.neutron [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 602.757000] env[62753]: DEBUG nova.network.neutron [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 602.872916] env[62753]: DEBUG nova.compute.manager [req-f5eccaad-ba97-4198-aac9-2962f2e79089 req-c3c68dff-6820-4c14-b81f-5b02a3ca1515 service nova] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Received event network-changed-08b654e9-1488-4940-b8c8-5528f7f78709 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 602.873140] env[62753]: DEBUG nova.compute.manager [req-f5eccaad-ba97-4198-aac9-2962f2e79089 req-c3c68dff-6820-4c14-b81f-5b02a3ca1515 service nova] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Refreshing instance network info cache due to event network-changed-08b654e9-1488-4940-b8c8-5528f7f78709. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 602.873338] env[62753]: DEBUG oslo_concurrency.lockutils [req-f5eccaad-ba97-4198-aac9-2962f2e79089 req-c3c68dff-6820-4c14-b81f-5b02a3ca1515 service nova] Acquiring lock "refresh_cache-579053e4-d16f-42dc-99c6-69c2c0d68040" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 603.057393] env[62753]: DEBUG nova.network.neutron [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.091192] env[62753]: DEBUG nova.compute.utils [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 603.098878] env[62753]: DEBUG nova.compute.manager [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 603.098878] env[62753]: DEBUG nova.network.neutron [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 603.182695] env[62753]: DEBUG nova.network.neutron [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 603.232342] env[62753]: DEBUG nova.policy [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '27a27d7d5034428887b4a487e81c3bf5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ee14ea30e3c94ddcaeed1ffbf32cde30', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 603.260350] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Releasing lock "refresh_cache-29dfb4db-c9ae-4a5c-8574-b27b13cdc83c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.260350] env[62753]: DEBUG nova.compute.manager [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 603.260350] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 603.261357] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4382bcd2-de58-4cf5-a55c-a3c1ce6dd0a1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.271835] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 603.272102] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9971a931-bd5a-450a-b5bb-9d1e1f701c95 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.281290] env[62753]: DEBUG oslo_vmware.api [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Waiting for the task: (returnval){ [ 603.281290] env[62753]: value = "task-1332084" [ 603.281290] env[62753]: _type = "Task" [ 603.281290] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.292919] env[62753]: DEBUG oslo_vmware.api [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332084, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 603.466294] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b29d06b-f3ef-49b9-a8ac-11fc7244babf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.474772] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a6bf3d9-a92c-415a-9814-d29e37cd51b0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.508952] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4cbd548-aea8-47a2-90e2-70181fbe8022 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.518259] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e027e3d-e078-4184-9684-b3fd222bdaf7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.536858] env[62753]: DEBUG nova.compute.provider_tree [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 603.599566] env[62753]: DEBUG nova.compute.manager [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 603.686992] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Releasing lock "refresh_cache-579053e4-d16f-42dc-99c6-69c2c0d68040" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 603.687477] env[62753]: DEBUG nova.compute.manager [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 603.687896] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 603.688039] env[62753]: DEBUG oslo_concurrency.lockutils [req-f5eccaad-ba97-4198-aac9-2962f2e79089 req-c3c68dff-6820-4c14-b81f-5b02a3ca1515 service nova] Acquired lock "refresh_cache-579053e4-d16f-42dc-99c6-69c2c0d68040" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 603.688300] env[62753]: DEBUG nova.network.neutron [req-f5eccaad-ba97-4198-aac9-2962f2e79089 req-c3c68dff-6820-4c14-b81f-5b02a3ca1515 service nova] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Refreshing network info cache for port 08b654e9-1488-4940-b8c8-5528f7f78709 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 603.689899] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-c4cde5e9-ba23-4ffc-9c63-a349596e880e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.701611] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a860af5e-48c3-4838-9089-367c39814936 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.727234] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 579053e4-d16f-42dc-99c6-69c2c0d68040 could not be found. [ 603.727655] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 603.727932] env[62753]: INFO nova.compute.manager [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Took 0.04 seconds to destroy the instance on the hypervisor. [ 603.728234] env[62753]: DEBUG oslo.service.loopingcall [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 603.728534] env[62753]: DEBUG nova.compute.manager [-] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 603.728724] env[62753]: DEBUG nova.network.neutron [-] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 603.792938] env[62753]: DEBUG oslo_vmware.api [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332084, 'name': PowerOffVM_Task, 'duration_secs': 0.124348} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 603.793701] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 603.793701] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 603.793839] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-de5a8715-d3fd-4962-9e89-f8a09d15c06e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.803742] env[62753]: DEBUG nova.network.neutron [-] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 603.819971] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 603.819971] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 603.819971] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Deleting the datastore file [datastore2] 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 603.824084] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0490db86-d8ad-4c9d-a9cf-9e0dda5264be {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 603.831306] env[62753]: DEBUG oslo_vmware.api [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Waiting for the task: (returnval){ [ 603.831306] env[62753]: value = "task-1332086" [ 603.831306] env[62753]: _type = "Task" [ 603.831306] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 603.841683] env[62753]: DEBUG oslo_vmware.api [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332086, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 604.040088] env[62753]: DEBUG nova.scheduler.client.report [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 604.127426] env[62753]: DEBUG nova.network.neutron [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Successfully created port: a78a6801-ac2e-44b6-9d0b-d48033df9eea {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 604.306945] env[62753]: DEBUG nova.network.neutron [-] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.343646] env[62753]: DEBUG nova.network.neutron [req-f5eccaad-ba97-4198-aac9-2962f2e79089 req-c3c68dff-6820-4c14-b81f-5b02a3ca1515 service nova] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.352153] env[62753]: DEBUG oslo_vmware.api [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Task: {'id': task-1332086, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.091176} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 604.352153] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 604.352153] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 604.352153] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 604.352153] env[62753]: INFO nova.compute.manager [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Took 1.09 seconds to destroy the instance on the hypervisor. [ 604.352388] env[62753]: DEBUG oslo.service.loopingcall [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 604.352388] env[62753]: DEBUG nova.compute.manager [-] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 604.352451] env[62753]: DEBUG nova.network.neutron [-] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 604.385211] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Acquiring lock "d81d5b85-fe31-43ec-91f6-7ff3251ff0c7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.385211] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Lock "d81d5b85-fe31-43ec-91f6-7ff3251ff0c7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.002s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.397957] env[62753]: DEBUG nova.network.neutron [-] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 604.547123] env[62753]: DEBUG nova.network.neutron [req-f5eccaad-ba97-4198-aac9-2962f2e79089 req-c3c68dff-6820-4c14-b81f-5b02a3ca1515 service nova] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.547123] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.963s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 604.547653] env[62753]: ERROR nova.compute.manager [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 1c21050c-9372-483e-9755-65a89cebd58f, please check neutron logs for more information. [ 604.547653] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Traceback (most recent call last): [ 604.547653] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 604.547653] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] self.driver.spawn(context, instance, image_meta, [ 604.547653] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 604.547653] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] self._vmops.spawn(context, instance, image_meta, injected_files, [ 604.547653] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 604.547653] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] vm_ref = self.build_virtual_machine(instance, [ 604.547653] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 604.547653] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] vif_infos = vmwarevif.get_vif_info(self._session, [ 604.547653] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 604.548099] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] for vif in network_info: [ 604.548099] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 604.548099] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] return self._sync_wrapper(fn, *args, **kwargs) [ 604.548099] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 604.548099] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] self.wait() [ 604.548099] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 604.548099] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] self[:] = self._gt.wait() [ 604.548099] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 604.548099] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] return self._exit_event.wait() [ 604.548099] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 604.548099] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] result = hub.switch() [ 604.548099] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 604.548099] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] return self.greenlet.switch() [ 604.548488] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 604.548488] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] result = function(*args, **kwargs) [ 604.548488] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 604.548488] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] return func(*args, **kwargs) [ 604.548488] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 604.548488] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] raise e [ 604.548488] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 604.548488] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] nwinfo = self.network_api.allocate_for_instance( [ 604.548488] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 604.548488] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] created_port_ids = self._update_ports_for_instance( [ 604.548488] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 604.548488] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] with excutils.save_and_reraise_exception(): [ 604.548488] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 604.548897] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] self.force_reraise() [ 604.548897] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 604.548897] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] raise self.value [ 604.548897] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 604.548897] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] updated_port = self._update_port( [ 604.548897] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 604.548897] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] _ensure_no_port_binding_failure(port) [ 604.548897] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 604.548897] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] raise exception.PortBindingFailed(port_id=port['id']) [ 604.548897] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] nova.exception.PortBindingFailed: Binding failed for port 1c21050c-9372-483e-9755-65a89cebd58f, please check neutron logs for more information. [ 604.548897] env[62753]: ERROR nova.compute.manager [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] [ 604.549675] env[62753]: DEBUG nova.compute.utils [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Binding failed for port 1c21050c-9372-483e-9755-65a89cebd58f, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 604.551674] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.449s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 604.554090] env[62753]: INFO nova.compute.claims [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 604.557977] env[62753]: DEBUG nova.compute.manager [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Build of instance 9afc71f1-21f3-422b-941f-23e29e29aa71 was re-scheduled: Binding failed for port 1c21050c-9372-483e-9755-65a89cebd58f, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 604.558839] env[62753]: DEBUG nova.compute.manager [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 604.559124] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Acquiring lock "refresh_cache-9afc71f1-21f3-422b-941f-23e29e29aa71" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 604.559290] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Acquired lock "refresh_cache-9afc71f1-21f3-422b-941f-23e29e29aa71" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 604.559569] env[62753]: DEBUG nova.network.neutron [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 604.610447] env[62753]: DEBUG nova.compute.manager [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 604.651536] env[62753]: DEBUG nova.virt.hardware [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 604.651859] env[62753]: DEBUG nova.virt.hardware [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 604.651948] env[62753]: DEBUG nova.virt.hardware [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 604.652143] env[62753]: DEBUG nova.virt.hardware [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 604.652273] env[62753]: DEBUG nova.virt.hardware [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 604.652416] env[62753]: DEBUG nova.virt.hardware [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 604.652624] env[62753]: DEBUG nova.virt.hardware [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 604.652782] env[62753]: DEBUG nova.virt.hardware [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 604.652948] env[62753]: DEBUG nova.virt.hardware [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 604.653228] env[62753]: DEBUG nova.virt.hardware [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 604.653422] env[62753]: DEBUG nova.virt.hardware [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 604.654331] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07b8abb9-c58b-4c00-98d3-b798fc5dd4f2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.662407] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70fd4856-9387-4a54-8782-a522516a571d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 604.813082] env[62753]: INFO nova.compute.manager [-] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Took 1.08 seconds to deallocate network for instance. [ 604.817497] env[62753]: DEBUG nova.compute.claims [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 604.817944] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.900641] env[62753]: DEBUG nova.network.neutron [-] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 604.934531] env[62753]: DEBUG oslo_concurrency.lockutils [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Acquiring lock "f10ff402-4308-4c75-af79-26e2942e0b92" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 604.935207] env[62753]: DEBUG oslo_concurrency.lockutils [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Lock "f10ff402-4308-4c75-af79-26e2942e0b92" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 605.052568] env[62753]: DEBUG oslo_concurrency.lockutils [req-f5eccaad-ba97-4198-aac9-2962f2e79089 req-c3c68dff-6820-4c14-b81f-5b02a3ca1515 service nova] Releasing lock "refresh_cache-579053e4-d16f-42dc-99c6-69c2c0d68040" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.089945] env[62753]: DEBUG nova.network.neutron [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.288673] env[62753]: DEBUG nova.compute.manager [req-9c94361e-462f-4a60-81ad-3fba3ef7e9ff req-1d582007-6b96-4c60-b0bd-7f20de41ac1b service nova] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Received event network-vif-deleted-08b654e9-1488-4940-b8c8-5528f7f78709 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 605.299514] env[62753]: DEBUG nova.network.neutron [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 605.404390] env[62753]: INFO nova.compute.manager [-] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Took 1.05 seconds to deallocate network for instance. [ 605.805632] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Releasing lock "refresh_cache-9afc71f1-21f3-422b-941f-23e29e29aa71" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 605.805851] env[62753]: DEBUG nova.compute.manager [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 605.806016] env[62753]: DEBUG nova.compute.manager [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 605.806237] env[62753]: DEBUG nova.network.neutron [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 605.903465] env[62753]: DEBUG nova.network.neutron [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 605.913062] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 606.021423] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a582bb5-9de7-4332-adf9-b383827c5f5c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.033225] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a5521544-84a4-4772-919e-b8a922020dcf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.066938] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f96d046c-48c8-4ff0-a9d6-094444b3d8d1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.074702] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d824306-274c-43ef-8f93-dc0ef2c9d349 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 606.092253] env[62753]: DEBUG nova.compute.provider_tree [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 606.205403] env[62753]: DEBUG nova.compute.manager [req-250485e1-ab9e-43dc-be78-69a00b6bb50e req-2dc8bc7c-93e9-4c11-99e7-086fed4a02ea service nova] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Received event network-changed-a78a6801-ac2e-44b6-9d0b-d48033df9eea {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 606.205589] env[62753]: DEBUG nova.compute.manager [req-250485e1-ab9e-43dc-be78-69a00b6bb50e req-2dc8bc7c-93e9-4c11-99e7-086fed4a02ea service nova] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Refreshing instance network info cache due to event network-changed-a78a6801-ac2e-44b6-9d0b-d48033df9eea. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 606.205854] env[62753]: DEBUG oslo_concurrency.lockutils [req-250485e1-ab9e-43dc-be78-69a00b6bb50e req-2dc8bc7c-93e9-4c11-99e7-086fed4a02ea service nova] Acquiring lock "refresh_cache-8d09404e-1359-42be-8795-226fcaadf0a3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.206169] env[62753]: DEBUG oslo_concurrency.lockutils [req-250485e1-ab9e-43dc-be78-69a00b6bb50e req-2dc8bc7c-93e9-4c11-99e7-086fed4a02ea service nova] Acquired lock "refresh_cache-8d09404e-1359-42be-8795-226fcaadf0a3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 606.206345] env[62753]: DEBUG nova.network.neutron [req-250485e1-ab9e-43dc-be78-69a00b6bb50e req-2dc8bc7c-93e9-4c11-99e7-086fed4a02ea service nova] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Refreshing network info cache for port a78a6801-ac2e-44b6-9d0b-d48033df9eea {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 606.346353] env[62753]: ERROR nova.compute.manager [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a78a6801-ac2e-44b6-9d0b-d48033df9eea, please check neutron logs for more information. [ 606.346353] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 606.346353] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.346353] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 606.346353] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.346353] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 606.346353] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.346353] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 606.346353] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.346353] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 606.346353] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.346353] env[62753]: ERROR nova.compute.manager raise self.value [ 606.346353] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.346353] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 606.346353] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.346353] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 606.346872] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.346872] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 606.346872] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a78a6801-ac2e-44b6-9d0b-d48033df9eea, please check neutron logs for more information. [ 606.346872] env[62753]: ERROR nova.compute.manager [ 606.346872] env[62753]: Traceback (most recent call last): [ 606.346872] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 606.346872] env[62753]: listener.cb(fileno) [ 606.346872] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.346872] env[62753]: result = function(*args, **kwargs) [ 606.346872] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 606.346872] env[62753]: return func(*args, **kwargs) [ 606.346872] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.346872] env[62753]: raise e [ 606.346872] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.346872] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 606.346872] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.346872] env[62753]: created_port_ids = self._update_ports_for_instance( [ 606.346872] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.346872] env[62753]: with excutils.save_and_reraise_exception(): [ 606.346872] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.346872] env[62753]: self.force_reraise() [ 606.346872] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.346872] env[62753]: raise self.value [ 606.346872] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.346872] env[62753]: updated_port = self._update_port( [ 606.346872] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.346872] env[62753]: _ensure_no_port_binding_failure(port) [ 606.346872] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.346872] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 606.348126] env[62753]: nova.exception.PortBindingFailed: Binding failed for port a78a6801-ac2e-44b6-9d0b-d48033df9eea, please check neutron logs for more information. [ 606.348126] env[62753]: Removing descriptor: 14 [ 606.348126] env[62753]: ERROR nova.compute.manager [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a78a6801-ac2e-44b6-9d0b-d48033df9eea, please check neutron logs for more information. [ 606.348126] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Traceback (most recent call last): [ 606.348126] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 606.348126] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] yield resources [ 606.348126] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 606.348126] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] self.driver.spawn(context, instance, image_meta, [ 606.348126] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 606.348126] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 606.348126] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 606.348126] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] vm_ref = self.build_virtual_machine(instance, [ 606.348728] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 606.348728] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] vif_infos = vmwarevif.get_vif_info(self._session, [ 606.348728] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 606.348728] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] for vif in network_info: [ 606.348728] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 606.348728] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] return self._sync_wrapper(fn, *args, **kwargs) [ 606.348728] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 606.348728] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] self.wait() [ 606.348728] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 606.348728] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] self[:] = self._gt.wait() [ 606.348728] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 606.348728] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] return self._exit_event.wait() [ 606.348728] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 606.349095] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] result = hub.switch() [ 606.349095] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 606.349095] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] return self.greenlet.switch() [ 606.349095] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 606.349095] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] result = function(*args, **kwargs) [ 606.349095] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 606.349095] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] return func(*args, **kwargs) [ 606.349095] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 606.349095] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] raise e [ 606.349095] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 606.349095] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] nwinfo = self.network_api.allocate_for_instance( [ 606.349095] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 606.349095] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] created_port_ids = self._update_ports_for_instance( [ 606.349648] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 606.349648] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] with excutils.save_and_reraise_exception(): [ 606.349648] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 606.349648] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] self.force_reraise() [ 606.349648] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 606.349648] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] raise self.value [ 606.349648] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 606.349648] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] updated_port = self._update_port( [ 606.349648] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 606.349648] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] _ensure_no_port_binding_failure(port) [ 606.349648] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 606.349648] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] raise exception.PortBindingFailed(port_id=port['id']) [ 606.350242] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] nova.exception.PortBindingFailed: Binding failed for port a78a6801-ac2e-44b6-9d0b-d48033df9eea, please check neutron logs for more information. [ 606.350242] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] [ 606.350242] env[62753]: INFO nova.compute.manager [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Terminating instance [ 606.350242] env[62753]: DEBUG oslo_concurrency.lockutils [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Acquiring lock "refresh_cache-8d09404e-1359-42be-8795-226fcaadf0a3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 606.408673] env[62753]: DEBUG nova.network.neutron [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 606.596187] env[62753]: DEBUG nova.scheduler.client.report [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 606.735500] env[62753]: DEBUG nova.network.neutron [req-250485e1-ab9e-43dc-be78-69a00b6bb50e req-2dc8bc7c-93e9-4c11-99e7-086fed4a02ea service nova] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 606.912590] env[62753]: INFO nova.compute.manager [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: 9afc71f1-21f3-422b-941f-23e29e29aa71] Took 1.11 seconds to deallocate network for instance. [ 606.978834] env[62753]: DEBUG nova.network.neutron [req-250485e1-ab9e-43dc-be78-69a00b6bb50e req-2dc8bc7c-93e9-4c11-99e7-086fed4a02ea service nova] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 607.101036] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.550s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 607.104181] env[62753]: DEBUG nova.compute.manager [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 607.108515] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.741s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 607.115034] env[62753]: INFO nova.compute.claims [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 607.482355] env[62753]: DEBUG oslo_concurrency.lockutils [req-250485e1-ab9e-43dc-be78-69a00b6bb50e req-2dc8bc7c-93e9-4c11-99e7-086fed4a02ea service nova] Releasing lock "refresh_cache-8d09404e-1359-42be-8795-226fcaadf0a3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 607.482355] env[62753]: DEBUG oslo_concurrency.lockutils [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Acquired lock "refresh_cache-8d09404e-1359-42be-8795-226fcaadf0a3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 607.482355] env[62753]: DEBUG nova.network.neutron [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 607.623316] env[62753]: DEBUG nova.compute.utils [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 607.630258] env[62753]: DEBUG nova.compute.manager [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 607.630452] env[62753]: DEBUG nova.network.neutron [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 607.752861] env[62753]: DEBUG nova.policy [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f0ca254418cf49ebb6e43ea4d38430b5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '9b70d6592502491c9be56f468fdd8509', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 607.963529] env[62753]: INFO nova.scheduler.client.report [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Deleted allocations for instance 9afc71f1-21f3-422b-941f-23e29e29aa71 [ 608.055590] env[62753]: DEBUG nova.network.neutron [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 608.131882] env[62753]: DEBUG nova.compute.manager [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 608.414966] env[62753]: DEBUG nova.network.neutron [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 608.475354] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4145301b-68f9-4ac4-b520-7a07ae1cf0cf tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Lock "9afc71f1-21f3-422b-941f-23e29e29aa71" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 28.763s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 608.547756] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e30049a-deb8-47b9-90ca-fe22f7e16640 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.558025] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a668684d-ffb1-4cd5-8d37-2e417fa9b442 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.603994] env[62753]: DEBUG nova.network.neutron [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Successfully created port: 6c41bbf1-15c1-46c8-97a3-fa41e7f6076f {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 608.610870] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3fe4cda-cf31-4b33-92d3-b2c91ea1bfa8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.617258] env[62753]: DEBUG nova.compute.manager [req-96152687-21aa-4b45-9069-9bd7e205f472 req-1ff44a27-312e-4b97-8a64-6e38cb8c8748 service nova] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Received event network-vif-deleted-a78a6801-ac2e-44b6-9d0b-d48033df9eea {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 608.627778] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-593172fe-4d82-43a3-91a0-ae0c100c46dd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.658400] env[62753]: DEBUG nova.compute.provider_tree [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 608.918577] env[62753]: DEBUG oslo_concurrency.lockutils [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Releasing lock "refresh_cache-8d09404e-1359-42be-8795-226fcaadf0a3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 608.918980] env[62753]: DEBUG nova.compute.manager [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 608.919203] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 608.919522] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8e9bc193-c06a-4e09-b73b-16e5abf08b7a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.931676] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd0ce334-3d02-4611-89b1-4b69518612af {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 608.956991] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 8d09404e-1359-42be-8795-226fcaadf0a3 could not be found. [ 608.956991] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 608.957139] env[62753]: INFO nova.compute.manager [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 608.957462] env[62753]: DEBUG oslo.service.loopingcall [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 608.957536] env[62753]: DEBUG nova.compute.manager [-] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 608.957613] env[62753]: DEBUG nova.network.neutron [-] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 608.976871] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 608.990859] env[62753]: DEBUG nova.network.neutron [-] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 609.160389] env[62753]: DEBUG nova.compute.manager [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 609.164159] env[62753]: DEBUG nova.scheduler.client.report [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 609.201239] env[62753]: DEBUG nova.virt.hardware [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 609.201515] env[62753]: DEBUG nova.virt.hardware [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 609.201671] env[62753]: DEBUG nova.virt.hardware [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 609.201842] env[62753]: DEBUG nova.virt.hardware [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 609.201980] env[62753]: DEBUG nova.virt.hardware [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 609.203555] env[62753]: DEBUG nova.virt.hardware [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 609.203816] env[62753]: DEBUG nova.virt.hardware [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 609.203980] env[62753]: DEBUG nova.virt.hardware [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 609.204182] env[62753]: DEBUG nova.virt.hardware [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 609.204347] env[62753]: DEBUG nova.virt.hardware [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 609.204519] env[62753]: DEBUG nova.virt.hardware [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 609.205730] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6296e39-141f-4cc5-94a4-7e2191b7f14b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.218503] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63be800e-379a-42d3-87d9-d4b591d16ce2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 609.496953] env[62753]: DEBUG nova.network.neutron [-] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 609.513205] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 609.669402] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.561s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 609.671545] env[62753]: DEBUG nova.compute.manager [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 609.672669] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 19.110s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 610.000824] env[62753]: INFO nova.compute.manager [-] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Took 1.04 seconds to deallocate network for instance. [ 610.007022] env[62753]: DEBUG nova.compute.claims [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 610.007022] env[62753]: DEBUG oslo_concurrency.lockutils [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 610.177454] env[62753]: DEBUG nova.compute.utils [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 610.185182] env[62753]: DEBUG nova.compute.manager [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 610.185182] env[62753]: DEBUG nova.network.neutron [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 610.313855] env[62753]: DEBUG nova.policy [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3329da92c5bf4f2c99085f5c5cae0dec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '6ee6c1b778984cb3821ed416bf36f3af', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 610.647614] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43454e2f-9b6a-4930-b759-a48ddb78c5d4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.656655] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04cba039-7672-48b4-8636-7713794078be {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.696045] env[62753]: DEBUG nova.compute.manager [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 610.700644] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42d72833-5794-4d9d-96c0-fac14df26330 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.709349] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5588b910-e82a-41e2-a0a8-e6d3e45d13f5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 610.723221] env[62753]: DEBUG nova.compute.provider_tree [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 611.227238] env[62753]: DEBUG nova.scheduler.client.report [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 611.328370] env[62753]: DEBUG nova.network.neutron [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Successfully created port: 5ec0fcb7-de7b-49ab-a3f8-6da7ee0167c6 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 611.712247] env[62753]: DEBUG nova.compute.manager [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 611.732664] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.060s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 611.733345] env[62753]: ERROR nova.compute.manager [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4f5c359b-daa7-4424-a6e2-e32a57c957f2, please check neutron logs for more information. [ 611.733345] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Traceback (most recent call last): [ 611.733345] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 611.733345] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] self.driver.spawn(context, instance, image_meta, [ 611.733345] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 611.733345] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] self._vmops.spawn(context, instance, image_meta, injected_files, [ 611.733345] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 611.733345] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] vm_ref = self.build_virtual_machine(instance, [ 611.733345] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 611.733345] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] vif_infos = vmwarevif.get_vif_info(self._session, [ 611.733345] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 611.733878] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] for vif in network_info: [ 611.733878] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 611.733878] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] return self._sync_wrapper(fn, *args, **kwargs) [ 611.733878] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 611.733878] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] self.wait() [ 611.733878] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 611.733878] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] self[:] = self._gt.wait() [ 611.733878] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 611.733878] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] return self._exit_event.wait() [ 611.733878] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 611.733878] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] result = hub.switch() [ 611.733878] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 611.733878] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] return self.greenlet.switch() [ 611.734676] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 611.734676] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] result = function(*args, **kwargs) [ 611.734676] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 611.734676] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] return func(*args, **kwargs) [ 611.734676] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 611.734676] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] raise e [ 611.734676] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 611.734676] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] nwinfo = self.network_api.allocate_for_instance( [ 611.734676] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 611.734676] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] created_port_ids = self._update_ports_for_instance( [ 611.734676] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 611.734676] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] with excutils.save_and_reraise_exception(): [ 611.734676] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 611.735181] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] self.force_reraise() [ 611.735181] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 611.735181] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] raise self.value [ 611.735181] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 611.735181] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] updated_port = self._update_port( [ 611.735181] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 611.735181] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] _ensure_no_port_binding_failure(port) [ 611.735181] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 611.735181] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] raise exception.PortBindingFailed(port_id=port['id']) [ 611.735181] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] nova.exception.PortBindingFailed: Binding failed for port 4f5c359b-daa7-4424-a6e2-e32a57c957f2, please check neutron logs for more information. [ 611.735181] env[62753]: ERROR nova.compute.manager [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] [ 611.735515] env[62753]: DEBUG nova.compute.utils [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Binding failed for port 4f5c359b-daa7-4424-a6e2-e32a57c957f2, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 611.735515] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.791s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 611.736870] env[62753]: INFO nova.compute.claims [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 611.739523] env[62753]: DEBUG nova.compute.manager [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Build of instance 609fa528-f73f-4dbc-83a4-d19bb3d7d452 was re-scheduled: Binding failed for port 4f5c359b-daa7-4424-a6e2-e32a57c957f2, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 611.739994] env[62753]: DEBUG nova.compute.manager [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 611.740318] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Acquiring lock "refresh_cache-609fa528-f73f-4dbc-83a4-d19bb3d7d452" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 611.740384] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Acquired lock "refresh_cache-609fa528-f73f-4dbc-83a4-d19bb3d7d452" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 611.740560] env[62753]: DEBUG nova.network.neutron [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 611.757804] env[62753]: DEBUG nova.virt.hardware [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 611.758359] env[62753]: DEBUG nova.virt.hardware [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 611.759265] env[62753]: DEBUG nova.virt.hardware [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 611.759265] env[62753]: DEBUG nova.virt.hardware [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 611.759265] env[62753]: DEBUG nova.virt.hardware [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 611.759942] env[62753]: DEBUG nova.virt.hardware [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 611.759942] env[62753]: DEBUG nova.virt.hardware [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 611.759942] env[62753]: DEBUG nova.virt.hardware [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 611.760167] env[62753]: DEBUG nova.virt.hardware [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 611.760306] env[62753]: DEBUG nova.virt.hardware [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 611.760477] env[62753]: DEBUG nova.virt.hardware [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 611.761978] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-782b916d-9e1c-49d4-bf45-e46a51b706ad {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 611.770122] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db36c8fd-61d8-4cb6-94a9-dfc687aa557f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 612.288229] env[62753]: DEBUG nova.network.neutron [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 612.439298] env[62753]: DEBUG nova.network.neutron [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 612.942456] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Releasing lock "refresh_cache-609fa528-f73f-4dbc-83a4-d19bb3d7d452" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 612.942764] env[62753]: DEBUG nova.compute.manager [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 612.942930] env[62753]: DEBUG nova.compute.manager [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 612.943112] env[62753]: DEBUG nova.network.neutron [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 613.003748] env[62753]: DEBUG nova.network.neutron [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.038133] env[62753]: ERROR nova.compute.manager [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 6c41bbf1-15c1-46c8-97a3-fa41e7f6076f, please check neutron logs for more information. [ 613.038133] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 613.038133] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.038133] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 613.038133] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 613.038133] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 613.038133] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 613.038133] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 613.038133] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.038133] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 613.038133] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.038133] env[62753]: ERROR nova.compute.manager raise self.value [ 613.038133] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 613.038133] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 613.038133] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.038133] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 613.038906] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.038906] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 613.038906] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 6c41bbf1-15c1-46c8-97a3-fa41e7f6076f, please check neutron logs for more information. [ 613.038906] env[62753]: ERROR nova.compute.manager [ 613.038906] env[62753]: Traceback (most recent call last): [ 613.038906] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 613.038906] env[62753]: listener.cb(fileno) [ 613.038906] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.038906] env[62753]: result = function(*args, **kwargs) [ 613.038906] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 613.038906] env[62753]: return func(*args, **kwargs) [ 613.038906] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.038906] env[62753]: raise e [ 613.038906] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.038906] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 613.038906] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 613.038906] env[62753]: created_port_ids = self._update_ports_for_instance( [ 613.038906] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 613.038906] env[62753]: with excutils.save_and_reraise_exception(): [ 613.038906] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.038906] env[62753]: self.force_reraise() [ 613.038906] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.038906] env[62753]: raise self.value [ 613.038906] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 613.038906] env[62753]: updated_port = self._update_port( [ 613.038906] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.038906] env[62753]: _ensure_no_port_binding_failure(port) [ 613.038906] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.038906] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 613.039768] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 6c41bbf1-15c1-46c8-97a3-fa41e7f6076f, please check neutron logs for more information. [ 613.039768] env[62753]: Removing descriptor: 14 [ 613.041680] env[62753]: ERROR nova.compute.manager [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 6c41bbf1-15c1-46c8-97a3-fa41e7f6076f, please check neutron logs for more information. [ 613.041680] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Traceback (most recent call last): [ 613.041680] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 613.041680] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] yield resources [ 613.041680] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 613.041680] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] self.driver.spawn(context, instance, image_meta, [ 613.041680] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 613.041680] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 613.041680] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 613.041680] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] vm_ref = self.build_virtual_machine(instance, [ 613.041680] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 613.042168] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] vif_infos = vmwarevif.get_vif_info(self._session, [ 613.042168] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 613.042168] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] for vif in network_info: [ 613.042168] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 613.042168] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] return self._sync_wrapper(fn, *args, **kwargs) [ 613.042168] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 613.042168] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] self.wait() [ 613.042168] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 613.042168] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] self[:] = self._gt.wait() [ 613.042168] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 613.042168] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] return self._exit_event.wait() [ 613.042168] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 613.042168] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] result = hub.switch() [ 613.042569] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 613.042569] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] return self.greenlet.switch() [ 613.042569] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 613.042569] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] result = function(*args, **kwargs) [ 613.042569] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 613.042569] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] return func(*args, **kwargs) [ 613.042569] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 613.042569] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] raise e [ 613.042569] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 613.042569] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] nwinfo = self.network_api.allocate_for_instance( [ 613.042569] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 613.042569] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] created_port_ids = self._update_ports_for_instance( [ 613.042569] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 613.042937] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] with excutils.save_and_reraise_exception(): [ 613.042937] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 613.042937] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] self.force_reraise() [ 613.042937] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 613.042937] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] raise self.value [ 613.042937] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 613.042937] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] updated_port = self._update_port( [ 613.042937] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 613.042937] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] _ensure_no_port_binding_failure(port) [ 613.042937] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 613.042937] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] raise exception.PortBindingFailed(port_id=port['id']) [ 613.042937] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] nova.exception.PortBindingFailed: Binding failed for port 6c41bbf1-15c1-46c8-97a3-fa41e7f6076f, please check neutron logs for more information. [ 613.042937] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] [ 613.043336] env[62753]: INFO nova.compute.manager [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Terminating instance [ 613.047862] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Acquiring lock "refresh_cache-4e1ae4a1-a06c-4afc-9c55-c9a728624ed9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.048271] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Acquired lock "refresh_cache-4e1ae4a1-a06c-4afc-9c55-c9a728624ed9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 613.048271] env[62753]: DEBUG nova.network.neutron [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 613.148645] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a2fb034-ba45-47a0-9804-4c12f8ceff47 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.156863] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf766da-10d1-44a2-8ab5-093b1755b8a5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.193541] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2f297cb-2b9e-45ac-b134-9733900a1a40 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.201761] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b3a752f-a15a-4e51-aa1c-53b5669408d9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 613.216548] env[62753]: DEBUG nova.compute.provider_tree [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 613.432631] env[62753]: DEBUG nova.compute.manager [req-bed6dc20-974c-434f-81ca-4c94980d6692 req-70bfd04a-b984-43cf-bfd7-05b81a771fd6 service nova] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Received event network-changed-6c41bbf1-15c1-46c8-97a3-fa41e7f6076f {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 613.432845] env[62753]: DEBUG nova.compute.manager [req-bed6dc20-974c-434f-81ca-4c94980d6692 req-70bfd04a-b984-43cf-bfd7-05b81a771fd6 service nova] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Refreshing instance network info cache due to event network-changed-6c41bbf1-15c1-46c8-97a3-fa41e7f6076f. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 613.433052] env[62753]: DEBUG oslo_concurrency.lockutils [req-bed6dc20-974c-434f-81ca-4c94980d6692 req-70bfd04a-b984-43cf-bfd7-05b81a771fd6 service nova] Acquiring lock "refresh_cache-4e1ae4a1-a06c-4afc-9c55-c9a728624ed9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 613.506284] env[62753]: DEBUG nova.network.neutron [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 613.590595] env[62753]: DEBUG nova.network.neutron [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 613.720637] env[62753]: DEBUG nova.scheduler.client.report [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 614.013489] env[62753]: INFO nova.compute.manager [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] [instance: 609fa528-f73f-4dbc-83a4-d19bb3d7d452] Took 1.07 seconds to deallocate network for instance. [ 614.023869] env[62753]: DEBUG nova.network.neutron [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 614.225310] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.490s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 614.225887] env[62753]: DEBUG nova.compute.manager [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 614.228446] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.771s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 614.230259] env[62753]: INFO nova.compute.claims [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 614.525096] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Releasing lock "refresh_cache-4e1ae4a1-a06c-4afc-9c55-c9a728624ed9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 614.529016] env[62753]: DEBUG nova.compute.manager [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 614.529016] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 614.529016] env[62753]: DEBUG oslo_concurrency.lockutils [req-bed6dc20-974c-434f-81ca-4c94980d6692 req-70bfd04a-b984-43cf-bfd7-05b81a771fd6 service nova] Acquired lock "refresh_cache-4e1ae4a1-a06c-4afc-9c55-c9a728624ed9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.529016] env[62753]: DEBUG nova.network.neutron [req-bed6dc20-974c-434f-81ca-4c94980d6692 req-70bfd04a-b984-43cf-bfd7-05b81a771fd6 service nova] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Refreshing network info cache for port 6c41bbf1-15c1-46c8-97a3-fa41e7f6076f {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 614.529016] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2b0ced9e-26fb-4bd6-b61e-65ed08d2a6eb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.541337] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb08fc36-59c3-4cf7-892d-624051d46b63 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 614.570204] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9 could not be found. [ 614.570204] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 614.570204] env[62753]: INFO nova.compute.manager [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Took 0.04 seconds to destroy the instance on the hypervisor. [ 614.570204] env[62753]: DEBUG oslo.service.loopingcall [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 614.570204] env[62753]: DEBUG nova.compute.manager [-] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 614.570204] env[62753]: DEBUG nova.network.neutron [-] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 614.622337] env[62753]: DEBUG nova.network.neutron [-] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 614.738785] env[62753]: DEBUG nova.compute.utils [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 614.740221] env[62753]: DEBUG nova.compute.manager [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 614.744019] env[62753]: DEBUG nova.network.neutron [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 614.932041] env[62753]: DEBUG nova.policy [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '5e310f46dd9540d1a4ed6ba33f127c6e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '05bee3612edd4b3f873026898352dd15', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 614.965239] env[62753]: ERROR nova.compute.manager [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 5ec0fcb7-de7b-49ab-a3f8-6da7ee0167c6, please check neutron logs for more information. [ 614.965239] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 614.965239] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.965239] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 614.965239] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.965239] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 614.965239] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.965239] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 614.965239] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.965239] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 614.965239] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.965239] env[62753]: ERROR nova.compute.manager raise self.value [ 614.965239] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.965239] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 614.965239] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.965239] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 614.965969] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.965969] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 614.965969] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 5ec0fcb7-de7b-49ab-a3f8-6da7ee0167c6, please check neutron logs for more information. [ 614.965969] env[62753]: ERROR nova.compute.manager [ 614.965969] env[62753]: Traceback (most recent call last): [ 614.965969] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 614.965969] env[62753]: listener.cb(fileno) [ 614.965969] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.965969] env[62753]: result = function(*args, **kwargs) [ 614.965969] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 614.965969] env[62753]: return func(*args, **kwargs) [ 614.965969] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.965969] env[62753]: raise e [ 614.965969] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.965969] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 614.965969] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.965969] env[62753]: created_port_ids = self._update_ports_for_instance( [ 614.965969] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.965969] env[62753]: with excutils.save_and_reraise_exception(): [ 614.965969] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.965969] env[62753]: self.force_reraise() [ 614.965969] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.965969] env[62753]: raise self.value [ 614.965969] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.965969] env[62753]: updated_port = self._update_port( [ 614.965969] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.965969] env[62753]: _ensure_no_port_binding_failure(port) [ 614.965969] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.965969] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 614.966859] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 5ec0fcb7-de7b-49ab-a3f8-6da7ee0167c6, please check neutron logs for more information. [ 614.966859] env[62753]: Removing descriptor: 16 [ 614.966859] env[62753]: ERROR nova.compute.manager [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 5ec0fcb7-de7b-49ab-a3f8-6da7ee0167c6, please check neutron logs for more information. [ 614.966859] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Traceback (most recent call last): [ 614.966859] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 614.966859] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] yield resources [ 614.966859] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 614.966859] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] self.driver.spawn(context, instance, image_meta, [ 614.966859] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 614.966859] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] self._vmops.spawn(context, instance, image_meta, injected_files, [ 614.966859] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 614.966859] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] vm_ref = self.build_virtual_machine(instance, [ 614.967358] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 614.967358] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] vif_infos = vmwarevif.get_vif_info(self._session, [ 614.967358] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 614.967358] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] for vif in network_info: [ 614.967358] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 614.967358] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] return self._sync_wrapper(fn, *args, **kwargs) [ 614.967358] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 614.967358] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] self.wait() [ 614.967358] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 614.967358] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] self[:] = self._gt.wait() [ 614.967358] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 614.967358] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] return self._exit_event.wait() [ 614.967358] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 614.967773] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] result = hub.switch() [ 614.967773] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 614.967773] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] return self.greenlet.switch() [ 614.967773] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 614.967773] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] result = function(*args, **kwargs) [ 614.967773] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 614.967773] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] return func(*args, **kwargs) [ 614.967773] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 614.967773] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] raise e [ 614.967773] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 614.967773] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] nwinfo = self.network_api.allocate_for_instance( [ 614.967773] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 614.967773] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] created_port_ids = self._update_ports_for_instance( [ 614.968137] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 614.968137] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] with excutils.save_and_reraise_exception(): [ 614.968137] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 614.968137] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] self.force_reraise() [ 614.968137] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 614.968137] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] raise self.value [ 614.968137] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 614.968137] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] updated_port = self._update_port( [ 614.968137] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 614.968137] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] _ensure_no_port_binding_failure(port) [ 614.968137] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 614.968137] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] raise exception.PortBindingFailed(port_id=port['id']) [ 614.968449] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] nova.exception.PortBindingFailed: Binding failed for port 5ec0fcb7-de7b-49ab-a3f8-6da7ee0167c6, please check neutron logs for more information. [ 614.968449] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] [ 614.968449] env[62753]: INFO nova.compute.manager [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Terminating instance [ 614.969769] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Acquiring lock "refresh_cache-1ba742c4-9fa7-431e-83eb-abd86477ac24" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 614.970961] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Acquired lock "refresh_cache-1ba742c4-9fa7-431e-83eb-abd86477ac24" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 614.970961] env[62753]: DEBUG nova.network.neutron [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 615.082033] env[62753]: DEBUG nova.network.neutron [req-bed6dc20-974c-434f-81ca-4c94980d6692 req-70bfd04a-b984-43cf-bfd7-05b81a771fd6 service nova] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.084929] env[62753]: INFO nova.scheduler.client.report [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Deleted allocations for instance 609fa528-f73f-4dbc-83a4-d19bb3d7d452 [ 615.130828] env[62753]: DEBUG nova.network.neutron [-] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.190414] env[62753]: DEBUG nova.compute.manager [req-0003abf3-e6c4-472c-8736-2ecebb7a9a6e req-97508fe4-ff70-49dd-b865-0547d6460743 service nova] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Received event network-changed-5ec0fcb7-de7b-49ab-a3f8-6da7ee0167c6 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 615.191234] env[62753]: DEBUG nova.compute.manager [req-0003abf3-e6c4-472c-8736-2ecebb7a9a6e req-97508fe4-ff70-49dd-b865-0547d6460743 service nova] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Refreshing instance network info cache due to event network-changed-5ec0fcb7-de7b-49ab-a3f8-6da7ee0167c6. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 615.191234] env[62753]: DEBUG oslo_concurrency.lockutils [req-0003abf3-e6c4-472c-8736-2ecebb7a9a6e req-97508fe4-ff70-49dd-b865-0547d6460743 service nova] Acquiring lock "refresh_cache-1ba742c4-9fa7-431e-83eb-abd86477ac24" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 615.249669] env[62753]: DEBUG nova.compute.manager [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 615.294093] env[62753]: DEBUG nova.network.neutron [req-bed6dc20-974c-434f-81ca-4c94980d6692 req-70bfd04a-b984-43cf-bfd7-05b81a771fd6 service nova] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.500327] env[62753]: DEBUG nova.network.neutron [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 615.604030] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa01c199-05f5-496b-a27e-dd7103b5758b tempest-ServerDiagnosticsTest-1650935954 tempest-ServerDiagnosticsTest-1650935954-project-member] Lock "609fa528-f73f-4dbc-83a4-d19bb3d7d452" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 34.985s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 615.632704] env[62753]: INFO nova.compute.manager [-] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Took 1.06 seconds to deallocate network for instance. [ 615.640537] env[62753]: DEBUG nova.compute.claims [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 615.640537] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 615.686208] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d4fd9f3-e9b4-4ce0-800d-8b5778990c49 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.696773] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aed7465-1a5f-4690-8e46-4d0a3d545c4c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.701978] env[62753]: DEBUG nova.network.neutron [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 615.736160] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60f8a5fc-881c-49cc-8fdb-80c5ed93664d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.746255] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa63b1c4-4ab8-410b-82c7-ca59ea7e6dca {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 615.762520] env[62753]: DEBUG nova.compute.provider_tree [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 615.801278] env[62753]: DEBUG oslo_concurrency.lockutils [req-bed6dc20-974c-434f-81ca-4c94980d6692 req-70bfd04a-b984-43cf-bfd7-05b81a771fd6 service nova] Releasing lock "refresh_cache-4e1ae4a1-a06c-4afc-9c55-c9a728624ed9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.107377] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 616.204756] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Releasing lock "refresh_cache-1ba742c4-9fa7-431e-83eb-abd86477ac24" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 616.205490] env[62753]: DEBUG nova.compute.manager [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 616.205490] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 616.205694] env[62753]: DEBUG oslo_concurrency.lockutils [req-0003abf3-e6c4-472c-8736-2ecebb7a9a6e req-97508fe4-ff70-49dd-b865-0547d6460743 service nova] Acquired lock "refresh_cache-1ba742c4-9fa7-431e-83eb-abd86477ac24" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 616.205869] env[62753]: DEBUG nova.network.neutron [req-0003abf3-e6c4-472c-8736-2ecebb7a9a6e req-97508fe4-ff70-49dd-b865-0547d6460743 service nova] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Refreshing network info cache for port 5ec0fcb7-de7b-49ab-a3f8-6da7ee0167c6 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 616.207860] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-37857132-77fe-4856-97d4-fdb3446bf96a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.219724] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-310b576f-c5c6-4eba-b34a-e4750f7cecb5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.248799] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 1ba742c4-9fa7-431e-83eb-abd86477ac24 could not be found. [ 616.249257] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 616.249257] env[62753]: INFO nova.compute.manager [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Took 0.04 seconds to destroy the instance on the hypervisor. [ 616.250772] env[62753]: DEBUG oslo.service.loopingcall [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 616.250772] env[62753]: DEBUG nova.compute.manager [-] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 616.250772] env[62753]: DEBUG nova.network.neutron [-] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 616.267834] env[62753]: DEBUG nova.compute.manager [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 616.270283] env[62753]: DEBUG nova.scheduler.client.report [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 616.304801] env[62753]: DEBUG nova.network.neutron [-] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.317110] env[62753]: DEBUG nova.virt.hardware [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 616.317110] env[62753]: DEBUG nova.virt.hardware [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 616.317110] env[62753]: DEBUG nova.virt.hardware [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 616.317401] env[62753]: DEBUG nova.virt.hardware [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 616.317401] env[62753]: DEBUG nova.virt.hardware [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 616.317401] env[62753]: DEBUG nova.virt.hardware [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 616.317599] env[62753]: DEBUG nova.virt.hardware [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 616.317599] env[62753]: DEBUG nova.virt.hardware [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 616.317758] env[62753]: DEBUG nova.virt.hardware [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 616.317916] env[62753]: DEBUG nova.virt.hardware [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 616.320203] env[62753]: DEBUG nova.virt.hardware [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 616.321109] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ed407fd-c5c0-4869-8a9e-13a6bb7c1a43 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.332356] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d245de6e-e0ca-4da7-915e-82f5e3c17b14 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 616.543797] env[62753]: DEBUG nova.compute.manager [req-68dc7cd4-4bb1-4a8c-a3b8-18d02cec4e03 req-77e3cf81-80d7-4640-a201-a3a73c8e36d8 service nova] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Received event network-vif-deleted-6c41bbf1-15c1-46c8-97a3-fa41e7f6076f {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 616.599786] env[62753]: DEBUG nova.network.neutron [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Successfully created port: 9234fb1c-56c7-46a2-85e8-3ec5486cac7a {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 616.644959] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 616.740439] env[62753]: DEBUG nova.network.neutron [req-0003abf3-e6c4-472c-8736-2ecebb7a9a6e req-97508fe4-ff70-49dd-b865-0547d6460743 service nova] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 616.776808] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.547s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 616.776808] env[62753]: DEBUG nova.compute.manager [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 616.780419] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.705s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 616.782874] env[62753]: INFO nova.compute.claims [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 616.811023] env[62753]: DEBUG nova.network.neutron [-] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 616.926193] env[62753]: DEBUG nova.network.neutron [req-0003abf3-e6c4-472c-8736-2ecebb7a9a6e req-97508fe4-ff70-49dd-b865-0547d6460743 service nova] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 617.288268] env[62753]: DEBUG nova.compute.utils [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 617.294650] env[62753]: DEBUG nova.compute.manager [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 617.294650] env[62753]: DEBUG nova.network.neutron [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 617.315589] env[62753]: INFO nova.compute.manager [-] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Took 1.06 seconds to deallocate network for instance. [ 617.316669] env[62753]: DEBUG nova.compute.claims [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 617.316861] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 617.415182] env[62753]: DEBUG nova.policy [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9161502f8ef248dc838ea5849b6b2df1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e5fec1826285462eb1b1cffe04756f2a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 617.429388] env[62753]: DEBUG oslo_concurrency.lockutils [req-0003abf3-e6c4-472c-8736-2ecebb7a9a6e req-97508fe4-ff70-49dd-b865-0547d6460743 service nova] Releasing lock "refresh_cache-1ba742c4-9fa7-431e-83eb-abd86477ac24" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 617.794679] env[62753]: DEBUG nova.compute.manager [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 618.086688] env[62753]: DEBUG nova.compute.manager [req-a8d9810b-22b1-41e3-b6ef-7f523421ca7e req-44f7dd4e-d535-412f-91e3-49072490a67c service nova] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Received event network-vif-deleted-5ec0fcb7-de7b-49ab-a3f8-6da7ee0167c6 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 618.115459] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ebc8547-6050-4cae-a654-adc34fa70b00 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.123438] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8ae8717-828c-4694-9bdb-fa78b5325644 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.153644] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-632a996f-0384-4724-bd43-207da8c11806 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.161372] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6094076d-321e-4573-8a8e-221dd2395b34 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.176345] env[62753]: DEBUG nova.compute.provider_tree [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 618.293650] env[62753]: DEBUG nova.network.neutron [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Successfully created port: 4f7cdd79-dbcd-4758-a853-c2f8f2591818 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 618.685268] env[62753]: DEBUG nova.scheduler.client.report [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 618.811853] env[62753]: DEBUG nova.compute.manager [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 618.844025] env[62753]: DEBUG nova.virt.hardware [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 618.844025] env[62753]: DEBUG nova.virt.hardware [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 618.844025] env[62753]: DEBUG nova.virt.hardware [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 618.844257] env[62753]: DEBUG nova.virt.hardware [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 618.844257] env[62753]: DEBUG nova.virt.hardware [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 618.844257] env[62753]: DEBUG nova.virt.hardware [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 618.844257] env[62753]: DEBUG nova.virt.hardware [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 618.844379] env[62753]: DEBUG nova.virt.hardware [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 618.844478] env[62753]: DEBUG nova.virt.hardware [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 618.844868] env[62753]: DEBUG nova.virt.hardware [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 618.844868] env[62753]: DEBUG nova.virt.hardware [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 618.845965] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87595f5e-88f1-42da-bb49-5f306c94d083 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 618.856965] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07ab1231-773f-4bcf-ba95-8bb14acafc51 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 619.193098] env[62753]: ERROR nova.compute.manager [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 9234fb1c-56c7-46a2-85e8-3ec5486cac7a, please check neutron logs for more information. [ 619.193098] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 619.193098] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.193098] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 619.193098] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.193098] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 619.193098] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.193098] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 619.193098] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.193098] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 619.193098] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.193098] env[62753]: ERROR nova.compute.manager raise self.value [ 619.193098] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.193098] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 619.193098] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.193098] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 619.193836] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.193836] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 619.193836] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 9234fb1c-56c7-46a2-85e8-3ec5486cac7a, please check neutron logs for more information. [ 619.193836] env[62753]: ERROR nova.compute.manager [ 619.193836] env[62753]: Traceback (most recent call last): [ 619.193836] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 619.193836] env[62753]: listener.cb(fileno) [ 619.193836] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.193836] env[62753]: result = function(*args, **kwargs) [ 619.193836] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 619.193836] env[62753]: return func(*args, **kwargs) [ 619.193836] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.193836] env[62753]: raise e [ 619.193836] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.193836] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 619.193836] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.193836] env[62753]: created_port_ids = self._update_ports_for_instance( [ 619.193836] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.193836] env[62753]: with excutils.save_and_reraise_exception(): [ 619.193836] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.193836] env[62753]: self.force_reraise() [ 619.193836] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.193836] env[62753]: raise self.value [ 619.193836] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.193836] env[62753]: updated_port = self._update_port( [ 619.193836] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.193836] env[62753]: _ensure_no_port_binding_failure(port) [ 619.193836] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.193836] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 619.196173] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 9234fb1c-56c7-46a2-85e8-3ec5486cac7a, please check neutron logs for more information. [ 619.196173] env[62753]: Removing descriptor: 14 [ 619.196173] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.415s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 619.196173] env[62753]: DEBUG nova.compute.manager [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 619.200271] env[62753]: ERROR nova.compute.manager [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 9234fb1c-56c7-46a2-85e8-3ec5486cac7a, please check neutron logs for more information. [ 619.200271] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Traceback (most recent call last): [ 619.200271] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 619.200271] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] yield resources [ 619.200271] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 619.200271] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] self.driver.spawn(context, instance, image_meta, [ 619.200271] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 619.200271] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 619.200271] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 619.200271] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] vm_ref = self.build_virtual_machine(instance, [ 619.200271] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 619.200658] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] vif_infos = vmwarevif.get_vif_info(self._session, [ 619.200658] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 619.200658] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] for vif in network_info: [ 619.200658] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 619.200658] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] return self._sync_wrapper(fn, *args, **kwargs) [ 619.200658] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 619.200658] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] self.wait() [ 619.200658] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 619.200658] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] self[:] = self._gt.wait() [ 619.200658] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 619.200658] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] return self._exit_event.wait() [ 619.200658] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 619.200658] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] result = hub.switch() [ 619.201035] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 619.201035] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] return self.greenlet.switch() [ 619.201035] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 619.201035] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] result = function(*args, **kwargs) [ 619.201035] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 619.201035] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] return func(*args, **kwargs) [ 619.201035] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 619.201035] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] raise e [ 619.201035] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 619.201035] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] nwinfo = self.network_api.allocate_for_instance( [ 619.201035] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 619.201035] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] created_port_ids = self._update_ports_for_instance( [ 619.201035] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 619.201408] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] with excutils.save_and_reraise_exception(): [ 619.201408] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 619.201408] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] self.force_reraise() [ 619.201408] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 619.201408] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] raise self.value [ 619.201408] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 619.201408] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] updated_port = self._update_port( [ 619.201408] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 619.201408] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] _ensure_no_port_binding_failure(port) [ 619.201408] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 619.201408] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] raise exception.PortBindingFailed(port_id=port['id']) [ 619.201408] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] nova.exception.PortBindingFailed: Binding failed for port 9234fb1c-56c7-46a2-85e8-3ec5486cac7a, please check neutron logs for more information. [ 619.201408] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] [ 619.201883] env[62753]: INFO nova.compute.manager [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Terminating instance [ 619.201883] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 22.434s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 619.205853] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Acquiring lock "refresh_cache-3cd890d9-4f45-404d-ba3e-7104abcf2b7d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.205989] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Acquired lock "refresh_cache-3cd890d9-4f45-404d-ba3e-7104abcf2b7d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 619.206975] env[62753]: DEBUG nova.network.neutron [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 619.659342] env[62753]: DEBUG nova.compute.manager [req-42bd1948-18b4-46c7-a35a-d4e81600380f req-337b93d9-9e1c-45c3-99ca-c52d011d440a service nova] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Received event network-changed-9234fb1c-56c7-46a2-85e8-3ec5486cac7a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 619.659646] env[62753]: DEBUG nova.compute.manager [req-42bd1948-18b4-46c7-a35a-d4e81600380f req-337b93d9-9e1c-45c3-99ca-c52d011d440a service nova] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Refreshing instance network info cache due to event network-changed-9234fb1c-56c7-46a2-85e8-3ec5486cac7a. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 619.659705] env[62753]: DEBUG oslo_concurrency.lockutils [req-42bd1948-18b4-46c7-a35a-d4e81600380f req-337b93d9-9e1c-45c3-99ca-c52d011d440a service nova] Acquiring lock "refresh_cache-3cd890d9-4f45-404d-ba3e-7104abcf2b7d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 619.702269] env[62753]: DEBUG nova.compute.utils [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 619.703941] env[62753]: DEBUG nova.compute.manager [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 619.703941] env[62753]: DEBUG nova.network.neutron [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 619.780312] env[62753]: DEBUG nova.network.neutron [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 619.891784] env[62753]: DEBUG nova.policy [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9161502f8ef248dc838ea5849b6b2df1', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e5fec1826285462eb1b1cffe04756f2a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 619.926482] env[62753]: DEBUG oslo_concurrency.lockutils [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Acquiring lock "5fcbd37f-0568-43db-ae44-596eb4467391" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 619.926796] env[62753]: DEBUG oslo_concurrency.lockutils [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Lock "5fcbd37f-0568-43db-ae44-596eb4467391" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 620.181280] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4874bcfe-236a-431e-b288-457e74400208 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.194142] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef1993fa-791a-4293-8864-0fcc0e7a3fe0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.236024] env[62753]: DEBUG nova.compute.manager [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 620.238572] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c8d9516-dac1-417d-bd9f-69e2c2ba3ad2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.248479] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e403db-323c-4cb4-99b7-46fb12e01f6b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.266554] env[62753]: DEBUG nova.compute.provider_tree [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 620.302844] env[62753]: DEBUG nova.network.neutron [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 620.771264] env[62753]: DEBUG nova.scheduler.client.report [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 620.805875] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Releasing lock "refresh_cache-3cd890d9-4f45-404d-ba3e-7104abcf2b7d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 620.806388] env[62753]: DEBUG nova.compute.manager [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 620.806509] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 620.806764] env[62753]: DEBUG oslo_concurrency.lockutils [req-42bd1948-18b4-46c7-a35a-d4e81600380f req-337b93d9-9e1c-45c3-99ca-c52d011d440a service nova] Acquired lock "refresh_cache-3cd890d9-4f45-404d-ba3e-7104abcf2b7d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 620.806933] env[62753]: DEBUG nova.network.neutron [req-42bd1948-18b4-46c7-a35a-d4e81600380f req-337b93d9-9e1c-45c3-99ca-c52d011d440a service nova] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Refreshing network info cache for port 9234fb1c-56c7-46a2-85e8-3ec5486cac7a {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 620.812781] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-7baa3528-c973-4ea8-b9e9-ed5724eec9b8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.825163] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ff0c3d-9b67-4141-948f-e759a8ec3e76 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 620.849561] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3cd890d9-4f45-404d-ba3e-7104abcf2b7d could not be found. [ 620.849561] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 620.849744] env[62753]: INFO nova.compute.manager [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Took 0.04 seconds to destroy the instance on the hypervisor. [ 620.849911] env[62753]: DEBUG oslo.service.loopingcall [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 620.851032] env[62753]: DEBUG nova.compute.manager [-] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 620.851032] env[62753]: DEBUG nova.network.neutron [-] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 620.878696] env[62753]: DEBUG nova.network.neutron [-] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 620.880862] env[62753]: DEBUG nova.network.neutron [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Successfully created port: 99c15a1c-c10e-4c95-bf5f-695f7631aef8 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 621.249986] env[62753]: DEBUG nova.compute.manager [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 621.272852] env[62753]: DEBUG nova.virt.hardware [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 621.273113] env[62753]: DEBUG nova.virt.hardware [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 621.273279] env[62753]: DEBUG nova.virt.hardware [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 621.273513] env[62753]: DEBUG nova.virt.hardware [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 621.273664] env[62753]: DEBUG nova.virt.hardware [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 621.273811] env[62753]: DEBUG nova.virt.hardware [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 621.274069] env[62753]: DEBUG nova.virt.hardware [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 621.274271] env[62753]: DEBUG nova.virt.hardware [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 621.274516] env[62753]: DEBUG nova.virt.hardware [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 621.274682] env[62753]: DEBUG nova.virt.hardware [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 621.274850] env[62753]: DEBUG nova.virt.hardware [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 621.275729] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eac859f3-bab8-4f83-96cc-09c793a368aa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.278860] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.078s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 621.279461] env[62753]: ERROR nova.compute.manager [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 10636605-321f-4ab2-9789-c183fe7592ff, please check neutron logs for more information. [ 621.279461] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Traceback (most recent call last): [ 621.279461] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.279461] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] self.driver.spawn(context, instance, image_meta, [ 621.279461] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 621.279461] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.279461] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.279461] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] vm_ref = self.build_virtual_machine(instance, [ 621.279461] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.279461] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.279461] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.279981] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] for vif in network_info: [ 621.279981] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 621.279981] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] return self._sync_wrapper(fn, *args, **kwargs) [ 621.279981] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 621.279981] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] self.wait() [ 621.279981] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 621.279981] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] self[:] = self._gt.wait() [ 621.279981] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.279981] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] return self._exit_event.wait() [ 621.279981] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 621.279981] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] result = hub.switch() [ 621.279981] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 621.279981] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] return self.greenlet.switch() [ 621.280666] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.280666] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] result = function(*args, **kwargs) [ 621.280666] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 621.280666] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] return func(*args, **kwargs) [ 621.280666] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.280666] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] raise e [ 621.280666] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.280666] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] nwinfo = self.network_api.allocate_for_instance( [ 621.280666] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.280666] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] created_port_ids = self._update_ports_for_instance( [ 621.280666] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.280666] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] with excutils.save_and_reraise_exception(): [ 621.280666] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.281191] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] self.force_reraise() [ 621.281191] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.281191] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] raise self.value [ 621.281191] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.281191] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] updated_port = self._update_port( [ 621.281191] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.281191] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] _ensure_no_port_binding_failure(port) [ 621.281191] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.281191] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] raise exception.PortBindingFailed(port_id=port['id']) [ 621.281191] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] nova.exception.PortBindingFailed: Binding failed for port 10636605-321f-4ab2-9789-c183fe7592ff, please check neutron logs for more information. [ 621.281191] env[62753]: ERROR nova.compute.manager [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] [ 621.281491] env[62753]: DEBUG nova.compute.utils [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Binding failed for port 10636605-321f-4ab2-9789-c183fe7592ff, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 621.281491] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.464s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 621.285076] env[62753]: DEBUG nova.compute.manager [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Build of instance a1df2c79-3afe-4bd7-89dd-c792a36bdca2 was re-scheduled: Binding failed for port 10636605-321f-4ab2-9789-c183fe7592ff, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 621.285523] env[62753]: DEBUG nova.compute.manager [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 621.288017] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Acquiring lock "refresh_cache-a1df2c79-3afe-4bd7-89dd-c792a36bdca2" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.288017] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Acquired lock "refresh_cache-a1df2c79-3afe-4bd7-89dd-c792a36bdca2" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.288017] env[62753]: DEBUG nova.network.neutron [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.290248] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfb7be13-3b20-4f4d-bb40-0a33730b615d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 621.369504] env[62753]: DEBUG nova.network.neutron [req-42bd1948-18b4-46c7-a35a-d4e81600380f req-337b93d9-9e1c-45c3-99ca-c52d011d440a service nova] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.384465] env[62753]: DEBUG nova.network.neutron [-] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.415084] env[62753]: ERROR nova.compute.manager [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4f7cdd79-dbcd-4758-a853-c2f8f2591818, please check neutron logs for more information. [ 621.415084] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 621.415084] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.415084] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 621.415084] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.415084] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 621.415084] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.415084] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 621.415084] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.415084] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 621.415084] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.415084] env[62753]: ERROR nova.compute.manager raise self.value [ 621.415084] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.415084] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 621.415084] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.415084] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 621.415633] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.415633] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 621.415633] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4f7cdd79-dbcd-4758-a853-c2f8f2591818, please check neutron logs for more information. [ 621.415633] env[62753]: ERROR nova.compute.manager [ 621.415633] env[62753]: Traceback (most recent call last): [ 621.415633] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 621.415633] env[62753]: listener.cb(fileno) [ 621.415633] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.415633] env[62753]: result = function(*args, **kwargs) [ 621.415633] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 621.415633] env[62753]: return func(*args, **kwargs) [ 621.415633] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.415633] env[62753]: raise e [ 621.415633] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.415633] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 621.415633] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.415633] env[62753]: created_port_ids = self._update_ports_for_instance( [ 621.415633] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.415633] env[62753]: with excutils.save_and_reraise_exception(): [ 621.415633] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.415633] env[62753]: self.force_reraise() [ 621.415633] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.415633] env[62753]: raise self.value [ 621.415633] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.415633] env[62753]: updated_port = self._update_port( [ 621.415633] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.415633] env[62753]: _ensure_no_port_binding_failure(port) [ 621.415633] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.415633] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 621.416811] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 4f7cdd79-dbcd-4758-a853-c2f8f2591818, please check neutron logs for more information. [ 621.416811] env[62753]: Removing descriptor: 16 [ 621.416811] env[62753]: ERROR nova.compute.manager [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4f7cdd79-dbcd-4758-a853-c2f8f2591818, please check neutron logs for more information. [ 621.416811] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Traceback (most recent call last): [ 621.416811] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 621.416811] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] yield resources [ 621.416811] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 621.416811] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] self.driver.spawn(context, instance, image_meta, [ 621.416811] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 621.416811] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] self._vmops.spawn(context, instance, image_meta, injected_files, [ 621.416811] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 621.416811] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] vm_ref = self.build_virtual_machine(instance, [ 621.417242] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 621.417242] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] vif_infos = vmwarevif.get_vif_info(self._session, [ 621.417242] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 621.417242] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] for vif in network_info: [ 621.417242] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 621.417242] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] return self._sync_wrapper(fn, *args, **kwargs) [ 621.417242] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 621.417242] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] self.wait() [ 621.417242] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 621.417242] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] self[:] = self._gt.wait() [ 621.417242] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 621.417242] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] return self._exit_event.wait() [ 621.417242] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 621.417691] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] result = hub.switch() [ 621.417691] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 621.417691] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] return self.greenlet.switch() [ 621.417691] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 621.417691] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] result = function(*args, **kwargs) [ 621.417691] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 621.417691] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] return func(*args, **kwargs) [ 621.417691] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 621.417691] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] raise e [ 621.417691] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 621.417691] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] nwinfo = self.network_api.allocate_for_instance( [ 621.417691] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 621.417691] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] created_port_ids = self._update_ports_for_instance( [ 621.418153] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 621.418153] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] with excutils.save_and_reraise_exception(): [ 621.418153] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 621.418153] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] self.force_reraise() [ 621.418153] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 621.418153] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] raise self.value [ 621.418153] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 621.418153] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] updated_port = self._update_port( [ 621.418153] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 621.418153] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] _ensure_no_port_binding_failure(port) [ 621.418153] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 621.418153] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] raise exception.PortBindingFailed(port_id=port['id']) [ 621.418712] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] nova.exception.PortBindingFailed: Binding failed for port 4f7cdd79-dbcd-4758-a853-c2f8f2591818, please check neutron logs for more information. [ 621.418712] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] [ 621.418712] env[62753]: INFO nova.compute.manager [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Terminating instance [ 621.418868] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Acquiring lock "refresh_cache-b4e2a5f2-7225-47fa-883b-d892c92e3949" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 621.418948] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Acquired lock "refresh_cache-b4e2a5f2-7225-47fa-883b-d892c92e3949" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 621.419142] env[62753]: DEBUG nova.network.neutron [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 621.506048] env[62753]: DEBUG nova.network.neutron [req-42bd1948-18b4-46c7-a35a-d4e81600380f req-337b93d9-9e1c-45c3-99ca-c52d011d440a service nova] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.812748] env[62753]: DEBUG nova.network.neutron [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.887076] env[62753]: INFO nova.compute.manager [-] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Took 1.04 seconds to deallocate network for instance. [ 621.894296] env[62753]: DEBUG nova.compute.claims [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 621.894503] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 621.946743] env[62753]: DEBUG nova.network.neutron [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 621.948054] env[62753]: DEBUG nova.network.neutron [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 621.988907] env[62753]: DEBUG nova.compute.manager [req-00349bba-ce6f-4968-a2b5-ced376bb7556 req-6de84c21-0bda-44e1-8b24-2ac58af4b7c1 service nova] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Received event network-changed-4f7cdd79-dbcd-4758-a853-c2f8f2591818 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 621.989117] env[62753]: DEBUG nova.compute.manager [req-00349bba-ce6f-4968-a2b5-ced376bb7556 req-6de84c21-0bda-44e1-8b24-2ac58af4b7c1 service nova] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Refreshing instance network info cache due to event network-changed-4f7cdd79-dbcd-4758-a853-c2f8f2591818. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 621.989305] env[62753]: DEBUG oslo_concurrency.lockutils [req-00349bba-ce6f-4968-a2b5-ced376bb7556 req-6de84c21-0bda-44e1-8b24-2ac58af4b7c1 service nova] Acquiring lock "refresh_cache-b4e2a5f2-7225-47fa-883b-d892c92e3949" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 622.008822] env[62753]: DEBUG oslo_concurrency.lockutils [req-42bd1948-18b4-46c7-a35a-d4e81600380f req-337b93d9-9e1c-45c3-99ca-c52d011d440a service nova] Releasing lock "refresh_cache-3cd890d9-4f45-404d-ba3e-7104abcf2b7d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.060031] env[62753]: DEBUG nova.network.neutron [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 622.188139] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6981a90f-9ca7-4eb4-ade1-7971fa7e1fff {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.195731] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da47397d-20c9-4922-84cc-65bad04350a0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.228159] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6e39459-ada1-44b0-b89d-1fac448fd3d7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.236255] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed3abec7-c72a-4182-a050-b32745187de6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.250519] env[62753]: DEBUG nova.compute.provider_tree [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 622.450698] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Releasing lock "refresh_cache-a1df2c79-3afe-4bd7-89dd-c792a36bdca2" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.450937] env[62753]: DEBUG nova.compute.manager [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 622.451139] env[62753]: DEBUG nova.compute.manager [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.451308] env[62753]: DEBUG nova.network.neutron [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.468570] env[62753]: DEBUG nova.network.neutron [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.563024] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Releasing lock "refresh_cache-b4e2a5f2-7225-47fa-883b-d892c92e3949" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 622.563452] env[62753]: DEBUG nova.compute.manager [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 622.563648] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 622.563988] env[62753]: DEBUG oslo_concurrency.lockutils [req-00349bba-ce6f-4968-a2b5-ced376bb7556 req-6de84c21-0bda-44e1-8b24-2ac58af4b7c1 service nova] Acquired lock "refresh_cache-b4e2a5f2-7225-47fa-883b-d892c92e3949" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 622.564181] env[62753]: DEBUG nova.network.neutron [req-00349bba-ce6f-4968-a2b5-ced376bb7556 req-6de84c21-0bda-44e1-8b24-2ac58af4b7c1 service nova] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Refreshing network info cache for port 4f7cdd79-dbcd-4758-a853-c2f8f2591818 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 622.565339] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f5b30b1d-1983-4ba9-837c-166d908cb179 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.575267] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc338359-a2f1-4f48-b86a-4a04509d137c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 622.600057] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b4e2a5f2-7225-47fa-883b-d892c92e3949 could not be found. [ 622.600057] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 622.600057] env[62753]: INFO nova.compute.manager [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Took 0.04 seconds to destroy the instance on the hypervisor. [ 622.600441] env[62753]: DEBUG oslo.service.loopingcall [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 622.600512] env[62753]: DEBUG nova.compute.manager [-] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 622.600580] env[62753]: DEBUG nova.network.neutron [-] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 622.629274] env[62753]: DEBUG oslo_concurrency.lockutils [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Acquiring lock "efaf8c7c-3f38-4a18-97e8-099da390c7d8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 622.629274] env[62753]: DEBUG oslo_concurrency.lockutils [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Lock "efaf8c7c-3f38-4a18-97e8-099da390c7d8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 622.650259] env[62753]: DEBUG nova.network.neutron [-] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 622.753436] env[62753]: DEBUG nova.scheduler.client.report [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 622.973480] env[62753]: DEBUG nova.network.neutron [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.091231] env[62753]: DEBUG nova.network.neutron [req-00349bba-ce6f-4968-a2b5-ced376bb7556 req-6de84c21-0bda-44e1-8b24-2ac58af4b7c1 service nova] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 623.156296] env[62753]: DEBUG nova.network.neutron [-] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.171404] env[62753]: DEBUG nova.network.neutron [req-00349bba-ce6f-4968-a2b5-ced376bb7556 req-6de84c21-0bda-44e1-8b24-2ac58af4b7c1 service nova] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 623.268734] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.987s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 623.275034] env[62753]: ERROR nova.compute.manager [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 08b654e9-1488-4940-b8c8-5528f7f78709, please check neutron logs for more information. [ 623.275034] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Traceback (most recent call last): [ 623.275034] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 623.275034] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] self.driver.spawn(context, instance, image_meta, [ 623.275034] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 623.275034] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.275034] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.275034] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] vm_ref = self.build_virtual_machine(instance, [ 623.275034] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.275034] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.275034] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.275847] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] for vif in network_info: [ 623.275847] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 623.275847] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] return self._sync_wrapper(fn, *args, **kwargs) [ 623.275847] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 623.275847] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] self.wait() [ 623.275847] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 623.275847] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] self[:] = self._gt.wait() [ 623.275847] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.275847] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] return self._exit_event.wait() [ 623.275847] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.275847] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] result = hub.switch() [ 623.275847] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.275847] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] return self.greenlet.switch() [ 623.276249] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.276249] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] result = function(*args, **kwargs) [ 623.276249] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 623.276249] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] return func(*args, **kwargs) [ 623.276249] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.276249] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] raise e [ 623.276249] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.276249] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] nwinfo = self.network_api.allocate_for_instance( [ 623.276249] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.276249] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] created_port_ids = self._update_ports_for_instance( [ 623.276249] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.276249] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] with excutils.save_and_reraise_exception(): [ 623.276249] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.276868] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] self.force_reraise() [ 623.276868] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.276868] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] raise self.value [ 623.276868] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.276868] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] updated_port = self._update_port( [ 623.276868] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.276868] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] _ensure_no_port_binding_failure(port) [ 623.276868] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.276868] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] raise exception.PortBindingFailed(port_id=port['id']) [ 623.276868] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] nova.exception.PortBindingFailed: Binding failed for port 08b654e9-1488-4940-b8c8-5528f7f78709, please check neutron logs for more information. [ 623.276868] env[62753]: ERROR nova.compute.manager [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] [ 623.277405] env[62753]: DEBUG nova.compute.utils [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Binding failed for port 08b654e9-1488-4940-b8c8-5528f7f78709, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 623.277405] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.362s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 623.277405] env[62753]: DEBUG nova.objects.instance [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Lazy-loading 'resources' on Instance uuid 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 623.280183] env[62753]: DEBUG nova.compute.manager [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Build of instance 579053e4-d16f-42dc-99c6-69c2c0d68040 was re-scheduled: Binding failed for port 08b654e9-1488-4940-b8c8-5528f7f78709, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 623.282520] env[62753]: DEBUG nova.compute.manager [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 623.282520] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Acquiring lock "refresh_cache-579053e4-d16f-42dc-99c6-69c2c0d68040" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.282520] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Acquired lock "refresh_cache-579053e4-d16f-42dc-99c6-69c2c0d68040" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.282520] env[62753]: DEBUG nova.network.neutron [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 623.394749] env[62753]: DEBUG nova.compute.manager [req-728962f5-ab8a-4741-83e0-303014c76135 req-8f74f2b1-44e7-404a-8234-81a0e6ae1d7f service nova] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Received event network-vif-deleted-9234fb1c-56c7-46a2-85e8-3ec5486cac7a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 623.477510] env[62753]: INFO nova.compute.manager [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] [instance: a1df2c79-3afe-4bd7-89dd-c792a36bdca2] Took 1.03 seconds to deallocate network for instance. [ 623.661964] env[62753]: INFO nova.compute.manager [-] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Took 1.06 seconds to deallocate network for instance. [ 623.662385] env[62753]: DEBUG nova.compute.claims [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 623.662518] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 623.682164] env[62753]: DEBUG oslo_concurrency.lockutils [req-00349bba-ce6f-4968-a2b5-ced376bb7556 req-6de84c21-0bda-44e1-8b24-2ac58af4b7c1 service nova] Releasing lock "refresh_cache-b4e2a5f2-7225-47fa-883b-d892c92e3949" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 623.701020] env[62753]: ERROR nova.compute.manager [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 99c15a1c-c10e-4c95-bf5f-695f7631aef8, please check neutron logs for more information. [ 623.701020] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 623.701020] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.701020] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 623.701020] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.701020] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 623.701020] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.701020] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 623.701020] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.701020] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 623.701020] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.701020] env[62753]: ERROR nova.compute.manager raise self.value [ 623.701020] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.701020] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 623.701020] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.701020] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 623.705094] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.705094] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 623.705094] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 99c15a1c-c10e-4c95-bf5f-695f7631aef8, please check neutron logs for more information. [ 623.705094] env[62753]: ERROR nova.compute.manager [ 623.705094] env[62753]: Traceback (most recent call last): [ 623.705094] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 623.705094] env[62753]: listener.cb(fileno) [ 623.705094] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.705094] env[62753]: result = function(*args, **kwargs) [ 623.705094] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 623.705094] env[62753]: return func(*args, **kwargs) [ 623.705094] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.705094] env[62753]: raise e [ 623.705094] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.705094] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 623.705094] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.705094] env[62753]: created_port_ids = self._update_ports_for_instance( [ 623.705094] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.705094] env[62753]: with excutils.save_and_reraise_exception(): [ 623.705094] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.705094] env[62753]: self.force_reraise() [ 623.705094] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.705094] env[62753]: raise self.value [ 623.705094] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.705094] env[62753]: updated_port = self._update_port( [ 623.705094] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.705094] env[62753]: _ensure_no_port_binding_failure(port) [ 623.705094] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.705094] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 623.706165] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 99c15a1c-c10e-4c95-bf5f-695f7631aef8, please check neutron logs for more information. [ 623.706165] env[62753]: Removing descriptor: 14 [ 623.706165] env[62753]: ERROR nova.compute.manager [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 99c15a1c-c10e-4c95-bf5f-695f7631aef8, please check neutron logs for more information. [ 623.706165] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Traceback (most recent call last): [ 623.706165] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 623.706165] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] yield resources [ 623.706165] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 623.706165] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] self.driver.spawn(context, instance, image_meta, [ 623.706165] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 623.706165] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 623.706165] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 623.706165] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] vm_ref = self.build_virtual_machine(instance, [ 623.706510] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 623.706510] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] vif_infos = vmwarevif.get_vif_info(self._session, [ 623.706510] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 623.706510] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] for vif in network_info: [ 623.706510] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 623.706510] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] return self._sync_wrapper(fn, *args, **kwargs) [ 623.706510] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 623.706510] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] self.wait() [ 623.706510] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 623.706510] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] self[:] = self._gt.wait() [ 623.706510] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 623.706510] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] return self._exit_event.wait() [ 623.706510] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 623.706863] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] result = hub.switch() [ 623.706863] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 623.706863] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] return self.greenlet.switch() [ 623.706863] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 623.706863] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] result = function(*args, **kwargs) [ 623.706863] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 623.706863] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] return func(*args, **kwargs) [ 623.706863] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 623.706863] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] raise e [ 623.706863] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 623.706863] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] nwinfo = self.network_api.allocate_for_instance( [ 623.706863] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 623.706863] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] created_port_ids = self._update_ports_for_instance( [ 623.710642] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 623.710642] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] with excutils.save_and_reraise_exception(): [ 623.710642] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 623.710642] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] self.force_reraise() [ 623.710642] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 623.710642] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] raise self.value [ 623.710642] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 623.710642] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] updated_port = self._update_port( [ 623.710642] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 623.710642] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] _ensure_no_port_binding_failure(port) [ 623.710642] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 623.710642] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] raise exception.PortBindingFailed(port_id=port['id']) [ 623.710968] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] nova.exception.PortBindingFailed: Binding failed for port 99c15a1c-c10e-4c95-bf5f-695f7631aef8, please check neutron logs for more information. [ 623.710968] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] [ 623.710968] env[62753]: INFO nova.compute.manager [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Terminating instance [ 623.710968] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Acquiring lock "refresh_cache-e5faf8e5-a5f6-4484-9304-7bbef03b33fc" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 623.710968] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Acquired lock "refresh_cache-e5faf8e5-a5f6-4484-9304-7bbef03b33fc" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 623.710968] env[62753]: DEBUG nova.network.neutron [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 623.827970] env[62753]: DEBUG nova.network.neutron [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.040105] env[62753]: DEBUG oslo_concurrency.lockutils [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "3f46fa44-4410-4ca9-98d0-8efa4cf06d52" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.040105] env[62753]: DEBUG oslo_concurrency.lockutils [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "3f46fa44-4410-4ca9-98d0-8efa4cf06d52" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.043512] env[62753]: DEBUG nova.network.neutron [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.242411] env[62753]: DEBUG nova.network.neutron [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.263279] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63734a91-15b2-4b60-8e3f-a441a3b37a76 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.273423] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d85571-3418-4438-91d8-607753f73d30 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.312611] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d75fdd-4549-4384-8de0-555d70ef17b4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.320654] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-044e1e5b-6eca-435d-828e-2c7d68d59b53 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.328474] env[62753]: DEBUG nova.network.neutron [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 624.332290] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Acquiring lock "5ef5c03c-6c22-4db6-95ad-6c407c0bca78" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.333018] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Lock "5ef5c03c-6c22-4db6-95ad-6c407c0bca78" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.343621] env[62753]: DEBUG nova.compute.provider_tree [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 624.434835] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Acquiring lock "50a480ed-f320-45d2-8ac9-a42d01614d11" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 624.435371] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Lock "50a480ed-f320-45d2-8ac9-a42d01614d11" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 624.532080] env[62753]: INFO nova.scheduler.client.report [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Deleted allocations for instance a1df2c79-3afe-4bd7-89dd-c792a36bdca2 [ 624.547146] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Releasing lock "refresh_cache-579053e4-d16f-42dc-99c6-69c2c0d68040" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.547146] env[62753]: DEBUG nova.compute.manager [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 624.547146] env[62753]: DEBUG nova.compute.manager [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 624.547146] env[62753]: DEBUG nova.network.neutron [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 624.570380] env[62753]: DEBUG nova.network.neutron [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 624.836033] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Releasing lock "refresh_cache-e5faf8e5-a5f6-4484-9304-7bbef03b33fc" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 624.836033] env[62753]: DEBUG nova.compute.manager [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 624.836033] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 624.836033] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-91523b20-fde8-465c-b81c-88324364e588 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.847031] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74984778-187d-4922-b188-1c7226475fe6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 624.862127] env[62753]: DEBUG nova.scheduler.client.report [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 624.865736] env[62753]: DEBUG nova.compute.manager [req-5f22902f-35a7-418f-b00c-e3a495736f21 req-23d04030-2b70-4cd4-a879-b528a12df9ce service nova] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Received event network-vif-deleted-4f7cdd79-dbcd-4758-a853-c2f8f2591818 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 624.879373] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e5faf8e5-a5f6-4484-9304-7bbef03b33fc could not be found. [ 624.879835] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 624.880567] env[62753]: INFO nova.compute.manager [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Took 0.05 seconds to destroy the instance on the hypervisor. [ 624.880567] env[62753]: DEBUG oslo.service.loopingcall [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 624.880567] env[62753]: DEBUG nova.compute.manager [-] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 624.880567] env[62753]: DEBUG nova.network.neutron [-] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 624.903754] env[62753]: DEBUG nova.network.neutron [-] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 625.042739] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84f4b9f8-94e0-4448-91be-241dfca24a7f tempest-ServersAdminTestJSON-1973766809 tempest-ServersAdminTestJSON-1973766809-project-member] Lock "a1df2c79-3afe-4bd7-89dd-c792a36bdca2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 41.846s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.074849] env[62753]: DEBUG nova.network.neutron [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.370387] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.096s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.373337] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.860s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 625.375486] env[62753]: INFO nova.compute.claims [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 625.399353] env[62753]: INFO nova.scheduler.client.report [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Deleted allocations for instance 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c [ 625.405791] env[62753]: DEBUG nova.network.neutron [-] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 625.546281] env[62753]: DEBUG nova.compute.manager [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 625.577346] env[62753]: INFO nova.compute.manager [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] [instance: 579053e4-d16f-42dc-99c6-69c2c0d68040] Took 1.03 seconds to deallocate network for instance. [ 625.741283] env[62753]: DEBUG nova.compute.manager [req-72253135-a0cc-490e-be1f-0a10703346df req-75d93e1a-ad2a-4ff7-9e14-16ddde34e07d service nova] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Received event network-changed-99c15a1c-c10e-4c95-bf5f-695f7631aef8 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 625.741283] env[62753]: DEBUG nova.compute.manager [req-72253135-a0cc-490e-be1f-0a10703346df req-75d93e1a-ad2a-4ff7-9e14-16ddde34e07d service nova] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Refreshing instance network info cache due to event network-changed-99c15a1c-c10e-4c95-bf5f-695f7631aef8. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 625.741414] env[62753]: DEBUG oslo_concurrency.lockutils [req-72253135-a0cc-490e-be1f-0a10703346df req-75d93e1a-ad2a-4ff7-9e14-16ddde34e07d service nova] Acquiring lock "refresh_cache-e5faf8e5-a5f6-4484-9304-7bbef03b33fc" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 625.744020] env[62753]: DEBUG oslo_concurrency.lockutils [req-72253135-a0cc-490e-be1f-0a10703346df req-75d93e1a-ad2a-4ff7-9e14-16ddde34e07d service nova] Acquired lock "refresh_cache-e5faf8e5-a5f6-4484-9304-7bbef03b33fc" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 625.744020] env[62753]: DEBUG nova.network.neutron [req-72253135-a0cc-490e-be1f-0a10703346df req-75d93e1a-ad2a-4ff7-9e14-16ddde34e07d service nova] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Refreshing network info cache for port 99c15a1c-c10e-4c95-bf5f-695f7631aef8 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 625.909394] env[62753]: INFO nova.compute.manager [-] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Took 1.03 seconds to deallocate network for instance. [ 625.917020] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a92b9589-2dfd-4843-ac60-5e09224339d6 tempest-ServerDiagnosticsV248Test-1751792613 tempest-ServerDiagnosticsV248Test-1751792613-project-member] Lock "29dfb4db-c9ae-4a5c-8574-b27b13cdc83c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.812s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 625.929523] env[62753]: DEBUG nova.compute.claims [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 625.930122] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.072452] env[62753]: DEBUG oslo_concurrency.lockutils [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 626.279593] env[62753]: DEBUG nova.network.neutron [req-72253135-a0cc-490e-be1f-0a10703346df req-75d93e1a-ad2a-4ff7-9e14-16ddde34e07d service nova] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 626.430091] env[62753]: DEBUG nova.network.neutron [req-72253135-a0cc-490e-be1f-0a10703346df req-75d93e1a-ad2a-4ff7-9e14-16ddde34e07d service nova] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 626.615765] env[62753]: INFO nova.scheduler.client.report [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Deleted allocations for instance 579053e4-d16f-42dc-99c6-69c2c0d68040 [ 626.825289] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d16d8a70-9a1f-49cb-9b23-80368f7cc238 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.835389] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c185585-345e-44a7-9028-37bee74bb585 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.881756] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59229af9-ab63-4326-ab2b-b7122d72a817 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.889809] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f171b82-cdba-4331-9fd8-14d2d515e3bb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 626.908405] env[62753]: DEBUG nova.compute.provider_tree [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 626.935368] env[62753]: DEBUG oslo_concurrency.lockutils [req-72253135-a0cc-490e-be1f-0a10703346df req-75d93e1a-ad2a-4ff7-9e14-16ddde34e07d service nova] Releasing lock "refresh_cache-e5faf8e5-a5f6-4484-9304-7bbef03b33fc" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 626.935635] env[62753]: DEBUG nova.compute.manager [req-72253135-a0cc-490e-be1f-0a10703346df req-75d93e1a-ad2a-4ff7-9e14-16ddde34e07d service nova] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Received event network-vif-deleted-99c15a1c-c10e-4c95-bf5f-695f7631aef8 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 627.128778] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1d8529ad-346f-4fcb-8704-0aad0c1c63a7 tempest-ImagesNegativeTestJSON-1413289370 tempest-ImagesNegativeTestJSON-1413289370-project-member] Lock "579053e4-d16f-42dc-99c6-69c2c0d68040" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 42.071s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.411812] env[62753]: DEBUG nova.scheduler.client.report [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 627.635395] env[62753]: DEBUG nova.compute.manager [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 627.917392] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.544s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 627.918448] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 627.922643] env[62753]: DEBUG oslo_concurrency.lockutils [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.918s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 628.164273] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 628.427465] env[62753]: DEBUG nova.compute.utils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 628.437889] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 628.437889] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 628.524218] env[62753]: DEBUG nova.policy [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbfa55fbe33b4312882c8a57a588f6b2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83648a9aa5354de3b0e43d03cea7c1f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 628.771553] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc1847c4-ab43-4605-8af9-e97a69e0faad {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.779040] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4acc6262-d735-4e0d-a00c-119d13b48678 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.813719] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9eaa9919-b655-4431-9aae-8894513f94b1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.822515] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67a6a67e-3e86-4339-bcf5-88146efa11c8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 628.838583] env[62753]: DEBUG nova.compute.provider_tree [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 628.936607] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 629.282751] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Successfully created port: 8f485a5c-58db-45f9-8fc7-928e5e7d8eb9 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 629.339872] env[62753]: DEBUG nova.scheduler.client.report [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 629.817462] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Acquiring lock "250f2711-6f32-4c4d-9876-35a3a1c30d12" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 629.817462] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Lock "250f2711-6f32-4c4d-9876-35a3a1c30d12" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.850164] env[62753]: DEBUG oslo_concurrency.lockutils [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.928s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 629.851208] env[62753]: ERROR nova.compute.manager [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a78a6801-ac2e-44b6-9d0b-d48033df9eea, please check neutron logs for more information. [ 629.851208] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Traceback (most recent call last): [ 629.851208] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 629.851208] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] self.driver.spawn(context, instance, image_meta, [ 629.851208] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 629.851208] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 629.851208] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 629.851208] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] vm_ref = self.build_virtual_machine(instance, [ 629.851208] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 629.851208] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] vif_infos = vmwarevif.get_vif_info(self._session, [ 629.851208] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 629.851984] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] for vif in network_info: [ 629.851984] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 629.851984] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] return self._sync_wrapper(fn, *args, **kwargs) [ 629.851984] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 629.851984] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] self.wait() [ 629.851984] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 629.851984] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] self[:] = self._gt.wait() [ 629.851984] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 629.851984] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] return self._exit_event.wait() [ 629.851984] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 629.851984] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] result = hub.switch() [ 629.851984] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 629.851984] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] return self.greenlet.switch() [ 629.853101] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 629.853101] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] result = function(*args, **kwargs) [ 629.853101] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 629.853101] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] return func(*args, **kwargs) [ 629.853101] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 629.853101] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] raise e [ 629.853101] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 629.853101] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] nwinfo = self.network_api.allocate_for_instance( [ 629.853101] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 629.853101] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] created_port_ids = self._update_ports_for_instance( [ 629.853101] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 629.853101] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] with excutils.save_and_reraise_exception(): [ 629.853101] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 629.854171] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] self.force_reraise() [ 629.854171] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 629.854171] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] raise self.value [ 629.854171] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 629.854171] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] updated_port = self._update_port( [ 629.854171] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 629.854171] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] _ensure_no_port_binding_failure(port) [ 629.854171] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 629.854171] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] raise exception.PortBindingFailed(port_id=port['id']) [ 629.854171] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] nova.exception.PortBindingFailed: Binding failed for port a78a6801-ac2e-44b6-9d0b-d48033df9eea, please check neutron logs for more information. [ 629.854171] env[62753]: ERROR nova.compute.manager [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] [ 629.855696] env[62753]: DEBUG nova.compute.utils [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Binding failed for port a78a6801-ac2e-44b6-9d0b-d48033df9eea, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 629.855696] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.215s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 629.862828] env[62753]: DEBUG nova.compute.manager [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Build of instance 8d09404e-1359-42be-8795-226fcaadf0a3 was re-scheduled: Binding failed for port a78a6801-ac2e-44b6-9d0b-d48033df9eea, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 629.863288] env[62753]: DEBUG nova.compute.manager [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 629.863505] env[62753]: DEBUG oslo_concurrency.lockutils [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Acquiring lock "refresh_cache-8d09404e-1359-42be-8795-226fcaadf0a3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 629.863662] env[62753]: DEBUG oslo_concurrency.lockutils [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Acquired lock "refresh_cache-8d09404e-1359-42be-8795-226fcaadf0a3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 629.863802] env[62753]: DEBUG nova.network.neutron [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 629.947782] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 629.987358] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 629.987358] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 629.987358] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 629.987553] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 629.987647] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 629.988127] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 629.994054] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 629.994054] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 629.994054] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 629.994054] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 629.994054] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 629.994352] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-685341ae-5506-4ff5-92d9-b335192983aa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.010080] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa4b0590-a597-4814-8bbf-4a2cddc777d7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.401670] env[62753]: DEBUG nova.network.neutron [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 630.489984] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Acquiring lock "a413174f-2a97-4c41-b512-e8ab562a25f4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.490916] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Lock "a413174f-2a97-4c41-b512-e8ab562a25f4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.579311] env[62753]: DEBUG nova.network.neutron [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 630.856604] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b6b33b6-ea65-4c90-bb8a-1f0e6263ac0e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.865805] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5a24979-d003-4820-b13b-173154cb2fce {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.896758] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19b6a207-08ee-4d88-9667-a09320d69148 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.904228] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c213a746-b9c2-4742-a9e7-37c158716a01 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 630.912340] env[62753]: DEBUG oslo_concurrency.lockutils [None req-67f953b7-d25d-4ea2-b4af-434ce4585a43 tempest-ServersListShow296Test-1505694806 tempest-ServersListShow296Test-1505694806-project-member] Acquiring lock "f0d83743-7a3b-4627-b69e-ac5c8e254a0e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 630.912544] env[62753]: DEBUG oslo_concurrency.lockutils [None req-67f953b7-d25d-4ea2-b4af-434ce4585a43 tempest-ServersListShow296Test-1505694806 tempest-ServersListShow296Test-1505694806-project-member] Lock "f0d83743-7a3b-4627-b69e-ac5c8e254a0e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 630.921778] env[62753]: DEBUG nova.compute.provider_tree [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 631.085522] env[62753]: DEBUG oslo_concurrency.lockutils [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Releasing lock "refresh_cache-8d09404e-1359-42be-8795-226fcaadf0a3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 631.085522] env[62753]: DEBUG nova.compute.manager [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 631.085522] env[62753]: DEBUG nova.compute.manager [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 631.085522] env[62753]: DEBUG nova.network.neutron [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 631.126590] env[62753]: DEBUG nova.network.neutron [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 631.427220] env[62753]: DEBUG nova.scheduler.client.report [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 631.544168] env[62753]: DEBUG nova.compute.manager [req-1b1aa79a-ddf3-4b36-9258-1b26b62a0c9c req-8f96f653-f799-4980-9bc1-74007c3b804d service nova] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Received event network-changed-8f485a5c-58db-45f9-8fc7-928e5e7d8eb9 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 631.544168] env[62753]: DEBUG nova.compute.manager [req-1b1aa79a-ddf3-4b36-9258-1b26b62a0c9c req-8f96f653-f799-4980-9bc1-74007c3b804d service nova] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Refreshing instance network info cache due to event network-changed-8f485a5c-58db-45f9-8fc7-928e5e7d8eb9. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 631.544168] env[62753]: DEBUG oslo_concurrency.lockutils [req-1b1aa79a-ddf3-4b36-9258-1b26b62a0c9c req-8f96f653-f799-4980-9bc1-74007c3b804d service nova] Acquiring lock "refresh_cache-11cf82b0-ea78-40b2-b737-dea8b8703ac3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.545058] env[62753]: DEBUG oslo_concurrency.lockutils [req-1b1aa79a-ddf3-4b36-9258-1b26b62a0c9c req-8f96f653-f799-4980-9bc1-74007c3b804d service nova] Acquired lock "refresh_cache-11cf82b0-ea78-40b2-b737-dea8b8703ac3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.545058] env[62753]: DEBUG nova.network.neutron [req-1b1aa79a-ddf3-4b36-9258-1b26b62a0c9c req-8f96f653-f799-4980-9bc1-74007c3b804d service nova] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Refreshing network info cache for port 8f485a5c-58db-45f9-8fc7-928e5e7d8eb9 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 631.631327] env[62753]: DEBUG nova.network.neutron [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 631.804116] env[62753]: ERROR nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8f485a5c-58db-45f9-8fc7-928e5e7d8eb9, please check neutron logs for more information. [ 631.804116] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 631.804116] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.804116] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 631.804116] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.804116] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 631.804116] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.804116] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 631.804116] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.804116] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 631.804116] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.804116] env[62753]: ERROR nova.compute.manager raise self.value [ 631.804116] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.804116] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 631.804116] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.804116] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 631.804744] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.804744] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 631.804744] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8f485a5c-58db-45f9-8fc7-928e5e7d8eb9, please check neutron logs for more information. [ 631.804744] env[62753]: ERROR nova.compute.manager [ 631.804744] env[62753]: Traceback (most recent call last): [ 631.804744] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 631.804744] env[62753]: listener.cb(fileno) [ 631.804744] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.804744] env[62753]: result = function(*args, **kwargs) [ 631.804744] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.804744] env[62753]: return func(*args, **kwargs) [ 631.804744] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.804744] env[62753]: raise e [ 631.804744] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.804744] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 631.804744] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.804744] env[62753]: created_port_ids = self._update_ports_for_instance( [ 631.804744] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.804744] env[62753]: with excutils.save_and_reraise_exception(): [ 631.804744] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.804744] env[62753]: self.force_reraise() [ 631.804744] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.804744] env[62753]: raise self.value [ 631.804744] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.804744] env[62753]: updated_port = self._update_port( [ 631.804744] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.804744] env[62753]: _ensure_no_port_binding_failure(port) [ 631.804744] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.804744] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 631.805731] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 8f485a5c-58db-45f9-8fc7-928e5e7d8eb9, please check neutron logs for more information. [ 631.805731] env[62753]: Removing descriptor: 14 [ 631.805731] env[62753]: ERROR nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8f485a5c-58db-45f9-8fc7-928e5e7d8eb9, please check neutron logs for more information. [ 631.805731] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Traceback (most recent call last): [ 631.805731] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 631.805731] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] yield resources [ 631.805731] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.805731] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] self.driver.spawn(context, instance, image_meta, [ 631.805731] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 631.805731] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.805731] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.805731] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] vm_ref = self.build_virtual_machine(instance, [ 631.806156] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.806156] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.806156] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.806156] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] for vif in network_info: [ 631.806156] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.806156] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] return self._sync_wrapper(fn, *args, **kwargs) [ 631.806156] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.806156] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] self.wait() [ 631.806156] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.806156] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] self[:] = self._gt.wait() [ 631.806156] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.806156] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] return self._exit_event.wait() [ 631.806156] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.806613] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] result = hub.switch() [ 631.806613] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.806613] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] return self.greenlet.switch() [ 631.806613] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.806613] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] result = function(*args, **kwargs) [ 631.806613] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.806613] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] return func(*args, **kwargs) [ 631.806613] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.806613] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] raise e [ 631.806613] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.806613] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] nwinfo = self.network_api.allocate_for_instance( [ 631.806613] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.806613] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] created_port_ids = self._update_ports_for_instance( [ 631.807024] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.807024] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] with excutils.save_and_reraise_exception(): [ 631.807024] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.807024] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] self.force_reraise() [ 631.807024] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.807024] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] raise self.value [ 631.807024] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.807024] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] updated_port = self._update_port( [ 631.807024] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.807024] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] _ensure_no_port_binding_failure(port) [ 631.807024] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.807024] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] raise exception.PortBindingFailed(port_id=port['id']) [ 631.808274] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] nova.exception.PortBindingFailed: Binding failed for port 8f485a5c-58db-45f9-8fc7-928e5e7d8eb9, please check neutron logs for more information. [ 631.808274] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] [ 631.808274] env[62753]: INFO nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Terminating instance [ 631.811261] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "refresh_cache-11cf82b0-ea78-40b2-b737-dea8b8703ac3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.935021] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.081s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 631.935021] env[62753]: ERROR nova.compute.manager [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 6c41bbf1-15c1-46c8-97a3-fa41e7f6076f, please check neutron logs for more information. [ 631.935021] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Traceback (most recent call last): [ 631.935021] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 631.935021] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] self.driver.spawn(context, instance, image_meta, [ 631.935021] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 631.935021] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 631.935021] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 631.935021] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] vm_ref = self.build_virtual_machine(instance, [ 631.935385] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 631.935385] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] vif_infos = vmwarevif.get_vif_info(self._session, [ 631.935385] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 631.935385] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] for vif in network_info: [ 631.935385] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 631.935385] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] return self._sync_wrapper(fn, *args, **kwargs) [ 631.935385] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 631.935385] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] self.wait() [ 631.935385] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 631.935385] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] self[:] = self._gt.wait() [ 631.935385] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 631.935385] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] return self._exit_event.wait() [ 631.935385] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 631.935746] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] result = hub.switch() [ 631.935746] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 631.935746] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] return self.greenlet.switch() [ 631.935746] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 631.935746] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] result = function(*args, **kwargs) [ 631.935746] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 631.935746] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] return func(*args, **kwargs) [ 631.935746] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 631.935746] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] raise e [ 631.935746] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 631.935746] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] nwinfo = self.network_api.allocate_for_instance( [ 631.935746] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 631.935746] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] created_port_ids = self._update_ports_for_instance( [ 631.936111] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 631.936111] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] with excutils.save_and_reraise_exception(): [ 631.936111] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 631.936111] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] self.force_reraise() [ 631.936111] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 631.936111] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] raise self.value [ 631.936111] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 631.936111] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] updated_port = self._update_port( [ 631.936111] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 631.936111] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] _ensure_no_port_binding_failure(port) [ 631.936111] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 631.936111] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] raise exception.PortBindingFailed(port_id=port['id']) [ 631.936444] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] nova.exception.PortBindingFailed: Binding failed for port 6c41bbf1-15c1-46c8-97a3-fa41e7f6076f, please check neutron logs for more information. [ 631.936444] env[62753]: ERROR nova.compute.manager [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] [ 631.936444] env[62753]: DEBUG nova.compute.utils [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Binding failed for port 6c41bbf1-15c1-46c8-97a3-fa41e7f6076f, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 631.938371] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.292s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 631.938492] env[62753]: INFO nova.compute.claims [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 631.941308] env[62753]: DEBUG nova.compute.manager [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Build of instance 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9 was re-scheduled: Binding failed for port 6c41bbf1-15c1-46c8-97a3-fa41e7f6076f, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 631.943322] env[62753]: DEBUG nova.compute.manager [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 631.943723] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Acquiring lock "refresh_cache-4e1ae4a1-a06c-4afc-9c55-c9a728624ed9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 631.943723] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Acquired lock "refresh_cache-4e1ae4a1-a06c-4afc-9c55-c9a728624ed9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 631.943842] env[62753]: DEBUG nova.network.neutron [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 632.081353] env[62753]: DEBUG nova.network.neutron [req-1b1aa79a-ddf3-4b36-9258-1b26b62a0c9c req-8f96f653-f799-4980-9bc1-74007c3b804d service nova] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.136327] env[62753]: INFO nova.compute.manager [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: 8d09404e-1359-42be-8795-226fcaadf0a3] Took 1.05 seconds to deallocate network for instance. [ 632.338730] env[62753]: DEBUG nova.network.neutron [req-1b1aa79a-ddf3-4b36-9258-1b26b62a0c9c req-8f96f653-f799-4980-9bc1-74007c3b804d service nova] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.481032] env[62753]: DEBUG nova.network.neutron [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 632.777888] env[62753]: DEBUG nova.network.neutron [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 632.842125] env[62753]: DEBUG oslo_concurrency.lockutils [req-1b1aa79a-ddf3-4b36-9258-1b26b62a0c9c req-8f96f653-f799-4980-9bc1-74007c3b804d service nova] Releasing lock "refresh_cache-11cf82b0-ea78-40b2-b737-dea8b8703ac3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 632.842388] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquired lock "refresh_cache-11cf82b0-ea78-40b2-b737-dea8b8703ac3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 632.842589] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 633.189122] env[62753]: INFO nova.scheduler.client.report [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Deleted allocations for instance 8d09404e-1359-42be-8795-226fcaadf0a3 [ 633.281695] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Releasing lock "refresh_cache-4e1ae4a1-a06c-4afc-9c55-c9a728624ed9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 633.281695] env[62753]: DEBUG nova.compute.manager [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 633.281695] env[62753]: DEBUG nova.compute.manager [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 633.281898] env[62753]: DEBUG nova.network.neutron [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 633.336706] env[62753]: DEBUG nova.network.neutron [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.394204] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 633.409776] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8cc9001-9065-4bf3-a67b-b7fed600147b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.419568] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa3d3664-7917-472a-99d3-be2b3e60fd5a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.449709] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df08fe4f-c49d-4fee-bd5e-f247e3f04983 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.457034] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21aa61bb-e560-42b1-a4f7-f5de7719abf2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 633.470746] env[62753]: DEBUG nova.compute.provider_tree [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 633.672341] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.699173] env[62753]: DEBUG oslo_concurrency.lockutils [None req-97e4e6ac-dd0f-44bf-b451-0a785e653853 tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Lock "8d09404e-1359-42be-8795-226fcaadf0a3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.465s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 633.843024] env[62753]: DEBUG nova.network.neutron [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 633.974582] env[62753]: DEBUG nova.scheduler.client.report [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 634.179893] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Releasing lock "refresh_cache-11cf82b0-ea78-40b2-b737-dea8b8703ac3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 634.181189] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 634.181189] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 634.181189] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9d38682b-edc1-4605-8707-10542a437736 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.193868] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca030a9c-94b1-4245-8983-ed6b7ff84d59 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 634.205430] env[62753]: DEBUG nova.compute.manager [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 634.222155] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 11cf82b0-ea78-40b2-b737-dea8b8703ac3 could not be found. [ 634.222390] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 634.222604] env[62753]: INFO nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Took 0.04 seconds to destroy the instance on the hypervisor. [ 634.222898] env[62753]: DEBUG oslo.service.loopingcall [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 634.223843] env[62753]: DEBUG nova.compute.manager [-] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 634.223843] env[62753]: DEBUG nova.network.neutron [-] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 634.245021] env[62753]: DEBUG nova.network.neutron [-] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 634.346249] env[62753]: INFO nova.compute.manager [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] [instance: 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9] Took 1.06 seconds to deallocate network for instance. [ 634.413481] env[62753]: DEBUG nova.compute.manager [req-eb1f8821-da8b-4938-a9b6-f2443e4ceaaf req-947cd6fd-c3a2-4b52-a622-40f7f5af51fd service nova] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Received event network-vif-deleted-8f485a5c-58db-45f9-8fc7-928e5e7d8eb9 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 634.480889] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.544s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 634.481388] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 634.485704] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.168s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.732828] env[62753]: DEBUG oslo_concurrency.lockutils [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.750079] env[62753]: DEBUG nova.network.neutron [-] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 634.914497] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "c4ccc036-4963-4499-8bbb-dad5c0e4687e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 634.914732] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "c4ccc036-4963-4499-8bbb-dad5c0e4687e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 634.994183] env[62753]: DEBUG nova.compute.utils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 634.994183] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 634.994183] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 635.134719] env[62753]: DEBUG nova.policy [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbfa55fbe33b4312882c8a57a588f6b2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83648a9aa5354de3b0e43d03cea7c1f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 635.251561] env[62753]: INFO nova.compute.manager [-] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Took 1.03 seconds to deallocate network for instance. [ 635.259106] env[62753]: DEBUG nova.compute.claims [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 635.259743] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.385632] env[62753]: INFO nova.scheduler.client.report [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Deleted allocations for instance 4e1ae4a1-a06c-4afc-9c55-c9a728624ed9 [ 635.407637] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-015f660f-3f11-4a8c-929e-55a9ec7fa9e6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.421530] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8366e674-fdd4-4ba8-9f1b-de7d94cb66c9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.463966] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9ecff8-0625-4cb8-a95f-b16c16aacb78 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.472911] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dfe06cf-f5d3-4095-bb5d-4f13437d7c3e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 635.487675] env[62753]: DEBUG nova.compute.provider_tree [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 635.503017] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 635.562312] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquiring lock "a8ceac07-6631-4ac6-b62a-7f3eac788f2b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 635.562545] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Lock "a8ceac07-6631-4ac6-b62a-7f3eac788f2b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 635.893922] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd05773a-edce-4f37-b1d6-6adfc329cda0 tempest-AttachInterfacesV270Test-769189834 tempest-AttachInterfacesV270Test-769189834-project-member] Lock "4e1ae4a1-a06c-4afc-9c55-c9a728624ed9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 47.822s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 635.992866] env[62753]: DEBUG nova.scheduler.client.report [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 636.039697] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Successfully created port: 75747bb4-bc24-4d44-812f-8f0a74badb68 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 636.396713] env[62753]: DEBUG nova.compute.manager [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 636.499083] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.014s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 636.501784] env[62753]: ERROR nova.compute.manager [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 5ec0fcb7-de7b-49ab-a3f8-6da7ee0167c6, please check neutron logs for more information. [ 636.501784] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Traceback (most recent call last): [ 636.501784] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 636.501784] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] self.driver.spawn(context, instance, image_meta, [ 636.501784] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 636.501784] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] self._vmops.spawn(context, instance, image_meta, injected_files, [ 636.501784] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 636.501784] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] vm_ref = self.build_virtual_machine(instance, [ 636.501784] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 636.501784] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] vif_infos = vmwarevif.get_vif_info(self._session, [ 636.501784] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 636.502450] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] for vif in network_info: [ 636.502450] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 636.502450] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] return self._sync_wrapper(fn, *args, **kwargs) [ 636.502450] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 636.502450] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] self.wait() [ 636.502450] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 636.502450] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] self[:] = self._gt.wait() [ 636.502450] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 636.502450] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] return self._exit_event.wait() [ 636.502450] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 636.502450] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] result = hub.switch() [ 636.502450] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 636.502450] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] return self.greenlet.switch() [ 636.502885] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 636.502885] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] result = function(*args, **kwargs) [ 636.502885] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 636.502885] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] return func(*args, **kwargs) [ 636.502885] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 636.502885] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] raise e [ 636.502885] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 636.502885] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] nwinfo = self.network_api.allocate_for_instance( [ 636.502885] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 636.502885] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] created_port_ids = self._update_ports_for_instance( [ 636.502885] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 636.502885] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] with excutils.save_and_reraise_exception(): [ 636.502885] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 636.503712] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] self.force_reraise() [ 636.503712] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 636.503712] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] raise self.value [ 636.503712] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 636.503712] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] updated_port = self._update_port( [ 636.503712] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 636.503712] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] _ensure_no_port_binding_failure(port) [ 636.503712] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 636.503712] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] raise exception.PortBindingFailed(port_id=port['id']) [ 636.503712] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] nova.exception.PortBindingFailed: Binding failed for port 5ec0fcb7-de7b-49ab-a3f8-6da7ee0167c6, please check neutron logs for more information. [ 636.503712] env[62753]: ERROR nova.compute.manager [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] [ 636.504078] env[62753]: DEBUG nova.compute.utils [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Binding failed for port 5ec0fcb7-de7b-49ab-a3f8-6da7ee0167c6, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 636.504078] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.609s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 636.509142] env[62753]: DEBUG nova.compute.manager [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Build of instance 1ba742c4-9fa7-431e-83eb-abd86477ac24 was re-scheduled: Binding failed for port 5ec0fcb7-de7b-49ab-a3f8-6da7ee0167c6, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 636.513209] env[62753]: DEBUG nova.compute.manager [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 636.513209] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Acquiring lock "refresh_cache-1ba742c4-9fa7-431e-83eb-abd86477ac24" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 636.513209] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Acquired lock "refresh_cache-1ba742c4-9fa7-431e-83eb-abd86477ac24" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 636.513209] env[62753]: DEBUG nova.network.neutron [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 636.513209] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 636.559072] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 636.559399] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 636.559982] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 636.560238] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 636.560461] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 636.560939] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 636.561208] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 636.561410] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 636.561642] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 636.561846] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 636.562067] env[62753]: DEBUG nova.virt.hardware [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 636.563277] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d720ab75-2525-4b19-b7bc-ca79b83773c9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.574353] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c89abe79-07b4-4a52-b449-18507ff6b111 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 636.922304] env[62753]: DEBUG oslo_concurrency.lockutils [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.048893] env[62753]: DEBUG nova.network.neutron [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 637.224662] env[62753]: DEBUG nova.network.neutron [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 637.361229] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Acquiring lock "da9b45e9-618d-4e1e-9c49-3d702b2435e9" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 637.361412] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Lock "da9b45e9-618d-4e1e-9c49-3d702b2435e9" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 637.443724] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e814b36-dc6d-4b9a-9dd0-523948ef2a13 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.451778] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3cc8513d-2090-405f-86ac-86ecdaf99532 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.487021] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9899f712-c9d3-4f96-b879-54edc778c984 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.492177] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9ba8fb-72be-44d1-9faa-86b0194a87a4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 637.511099] env[62753]: DEBUG nova.compute.provider_tree [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 637.729688] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Releasing lock "refresh_cache-1ba742c4-9fa7-431e-83eb-abd86477ac24" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 637.730011] env[62753]: DEBUG nova.compute.manager [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 637.730114] env[62753]: DEBUG nova.compute.manager [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 637.730301] env[62753]: DEBUG nova.network.neutron [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 637.754288] env[62753]: DEBUG nova.network.neutron [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 638.016540] env[62753]: DEBUG nova.scheduler.client.report [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 638.257223] env[62753]: DEBUG nova.network.neutron [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 638.520560] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.017s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 638.521223] env[62753]: ERROR nova.compute.manager [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 9234fb1c-56c7-46a2-85e8-3ec5486cac7a, please check neutron logs for more information. [ 638.521223] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Traceback (most recent call last): [ 638.521223] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 638.521223] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] self.driver.spawn(context, instance, image_meta, [ 638.521223] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 638.521223] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] self._vmops.spawn(context, instance, image_meta, injected_files, [ 638.521223] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 638.521223] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] vm_ref = self.build_virtual_machine(instance, [ 638.521223] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 638.521223] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] vif_infos = vmwarevif.get_vif_info(self._session, [ 638.521223] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 638.521640] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] for vif in network_info: [ 638.521640] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 638.521640] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] return self._sync_wrapper(fn, *args, **kwargs) [ 638.521640] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 638.521640] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] self.wait() [ 638.521640] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 638.521640] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] self[:] = self._gt.wait() [ 638.521640] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 638.521640] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] return self._exit_event.wait() [ 638.521640] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 638.521640] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] result = hub.switch() [ 638.521640] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 638.521640] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] return self.greenlet.switch() [ 638.522161] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 638.522161] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] result = function(*args, **kwargs) [ 638.522161] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 638.522161] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] return func(*args, **kwargs) [ 638.522161] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 638.522161] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] raise e [ 638.522161] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 638.522161] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] nwinfo = self.network_api.allocate_for_instance( [ 638.522161] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 638.522161] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] created_port_ids = self._update_ports_for_instance( [ 638.522161] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 638.522161] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] with excutils.save_and_reraise_exception(): [ 638.522161] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 638.522550] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] self.force_reraise() [ 638.522550] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 638.522550] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] raise self.value [ 638.522550] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 638.522550] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] updated_port = self._update_port( [ 638.522550] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 638.522550] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] _ensure_no_port_binding_failure(port) [ 638.522550] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 638.522550] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] raise exception.PortBindingFailed(port_id=port['id']) [ 638.522550] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] nova.exception.PortBindingFailed: Binding failed for port 9234fb1c-56c7-46a2-85e8-3ec5486cac7a, please check neutron logs for more information. [ 638.522550] env[62753]: ERROR nova.compute.manager [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] [ 638.522909] env[62753]: DEBUG nova.compute.utils [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Binding failed for port 9234fb1c-56c7-46a2-85e8-3ec5486cac7a, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 638.527019] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.861s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 638.532225] env[62753]: DEBUG nova.compute.manager [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Build of instance 3cd890d9-4f45-404d-ba3e-7104abcf2b7d was re-scheduled: Binding failed for port 9234fb1c-56c7-46a2-85e8-3ec5486cac7a, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 638.532759] env[62753]: DEBUG nova.compute.manager [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 638.533055] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Acquiring lock "refresh_cache-3cd890d9-4f45-404d-ba3e-7104abcf2b7d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 638.533225] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Acquired lock "refresh_cache-3cd890d9-4f45-404d-ba3e-7104abcf2b7d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 638.533386] env[62753]: DEBUG nova.network.neutron [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 638.765869] env[62753]: INFO nova.compute.manager [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] [instance: 1ba742c4-9fa7-431e-83eb-abd86477ac24] Took 1.04 seconds to deallocate network for instance. [ 639.010732] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Acquiring lock "e755857a-e0bc-40bf-bcbb-a416d243442b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.010854] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Lock "e755857a-e0bc-40bf-bcbb-a416d243442b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.068927] env[62753]: DEBUG nova.network.neutron [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.269116] env[62753]: DEBUG nova.network.neutron [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 639.383698] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Acquiring lock "eca070f0-dda0-4c51-b2d3-1251c517804a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 639.383830] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Lock "eca070f0-dda0-4c51-b2d3-1251c517804a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 639.467944] env[62753]: ERROR nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 75747bb4-bc24-4d44-812f-8f0a74badb68, please check neutron logs for more information. [ 639.467944] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 639.467944] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.467944] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 639.467944] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.467944] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 639.467944] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.467944] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 639.467944] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.467944] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 639.467944] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.467944] env[62753]: ERROR nova.compute.manager raise self.value [ 639.467944] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.467944] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 639.467944] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.467944] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 639.468503] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.468503] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 639.468503] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 75747bb4-bc24-4d44-812f-8f0a74badb68, please check neutron logs for more information. [ 639.468503] env[62753]: ERROR nova.compute.manager [ 639.469150] env[62753]: Traceback (most recent call last): [ 639.469150] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 639.469150] env[62753]: listener.cb(fileno) [ 639.469150] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.469150] env[62753]: result = function(*args, **kwargs) [ 639.469150] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 639.469150] env[62753]: return func(*args, **kwargs) [ 639.469150] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.469150] env[62753]: raise e [ 639.469150] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.469150] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 639.469150] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.469150] env[62753]: created_port_ids = self._update_ports_for_instance( [ 639.469150] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.469150] env[62753]: with excutils.save_and_reraise_exception(): [ 639.469150] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.469150] env[62753]: self.force_reraise() [ 639.469150] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.469150] env[62753]: raise self.value [ 639.469150] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.469150] env[62753]: updated_port = self._update_port( [ 639.469150] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.469150] env[62753]: _ensure_no_port_binding_failure(port) [ 639.469150] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.469150] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 639.469150] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 75747bb4-bc24-4d44-812f-8f0a74badb68, please check neutron logs for more information. [ 639.469150] env[62753]: Removing descriptor: 14 [ 639.470116] env[62753]: ERROR nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 75747bb4-bc24-4d44-812f-8f0a74badb68, please check neutron logs for more information. [ 639.470116] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Traceback (most recent call last): [ 639.470116] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 639.470116] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] yield resources [ 639.470116] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 639.470116] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] self.driver.spawn(context, instance, image_meta, [ 639.470116] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 639.470116] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 639.470116] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 639.470116] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] vm_ref = self.build_virtual_machine(instance, [ 639.470116] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 639.470513] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] vif_infos = vmwarevif.get_vif_info(self._session, [ 639.470513] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 639.470513] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] for vif in network_info: [ 639.470513] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 639.470513] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] return self._sync_wrapper(fn, *args, **kwargs) [ 639.470513] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 639.470513] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] self.wait() [ 639.470513] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 639.470513] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] self[:] = self._gt.wait() [ 639.470513] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 639.470513] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] return self._exit_event.wait() [ 639.470513] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 639.470513] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] result = hub.switch() [ 639.470965] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 639.470965] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] return self.greenlet.switch() [ 639.470965] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 639.470965] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] result = function(*args, **kwargs) [ 639.470965] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 639.470965] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] return func(*args, **kwargs) [ 639.470965] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 639.470965] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] raise e [ 639.470965] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 639.470965] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] nwinfo = self.network_api.allocate_for_instance( [ 639.470965] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 639.470965] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] created_port_ids = self._update_ports_for_instance( [ 639.470965] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 639.471473] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] with excutils.save_and_reraise_exception(): [ 639.471473] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 639.471473] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] self.force_reraise() [ 639.471473] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 639.471473] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] raise self.value [ 639.471473] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 639.471473] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] updated_port = self._update_port( [ 639.471473] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 639.471473] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] _ensure_no_port_binding_failure(port) [ 639.471473] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 639.471473] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] raise exception.PortBindingFailed(port_id=port['id']) [ 639.471473] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] nova.exception.PortBindingFailed: Binding failed for port 75747bb4-bc24-4d44-812f-8f0a74badb68, please check neutron logs for more information. [ 639.471473] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] [ 639.471855] env[62753]: INFO nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Terminating instance [ 639.473545] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "refresh_cache-32dae012-7743-4efb-b39b-820c9ed42cc5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.474594] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquired lock "refresh_cache-32dae012-7743-4efb-b39b-820c9ed42cc5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 639.474790] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 639.542643] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac61a43-07c0-4113-bb78-5982fee2593d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.555425] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b238c527-f9d9-4bbe-9276-e51aa57b3824 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.589724] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aa8bac3-9df5-4dd5-b336-a1dca55fbb78 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.594240] env[62753]: DEBUG nova.compute.manager [req-91c4312d-8fdc-45cd-b11c-d1853ec7c816 req-7d1dfd9b-9646-455c-92d3-0f70a4c478e6 service nova] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Received event network-changed-75747bb4-bc24-4d44-812f-8f0a74badb68 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 639.594386] env[62753]: DEBUG nova.compute.manager [req-91c4312d-8fdc-45cd-b11c-d1853ec7c816 req-7d1dfd9b-9646-455c-92d3-0f70a4c478e6 service nova] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Refreshing instance network info cache due to event network-changed-75747bb4-bc24-4d44-812f-8f0a74badb68. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 639.594589] env[62753]: DEBUG oslo_concurrency.lockutils [req-91c4312d-8fdc-45cd-b11c-d1853ec7c816 req-7d1dfd9b-9646-455c-92d3-0f70a4c478e6 service nova] Acquiring lock "refresh_cache-32dae012-7743-4efb-b39b-820c9ed42cc5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 639.601199] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246e18c5-25fd-43df-8d7d-f36fab4e631b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 639.614461] env[62753]: DEBUG nova.compute.provider_tree [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 639.775512] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Releasing lock "refresh_cache-3cd890d9-4f45-404d-ba3e-7104abcf2b7d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 639.777041] env[62753]: DEBUG nova.compute.manager [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 639.777041] env[62753]: DEBUG nova.compute.manager [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 639.777041] env[62753]: DEBUG nova.network.neutron [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 639.811980] env[62753]: DEBUG nova.network.neutron [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 639.816134] env[62753]: INFO nova.scheduler.client.report [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Deleted allocations for instance 1ba742c4-9fa7-431e-83eb-abd86477ac24 [ 640.022578] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.119208] env[62753]: DEBUG nova.scheduler.client.report [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 640.182421] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.324023] env[62753]: DEBUG nova.network.neutron [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 640.324023] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc1cfb8d-7566-4ec3-9182-80c82a23920a tempest-ServersTestJSON-1157657635 tempest-ServersTestJSON-1157657635-project-member] Lock "1ba742c4-9fa7-431e-83eb-abd86477ac24" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.981s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.628902] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.106s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 640.630460] env[62753]: ERROR nova.compute.manager [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4f7cdd79-dbcd-4758-a853-c2f8f2591818, please check neutron logs for more information. [ 640.630460] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Traceback (most recent call last): [ 640.630460] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 640.630460] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] self.driver.spawn(context, instance, image_meta, [ 640.630460] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 640.630460] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] self._vmops.spawn(context, instance, image_meta, injected_files, [ 640.630460] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 640.630460] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] vm_ref = self.build_virtual_machine(instance, [ 640.630460] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 640.630460] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] vif_infos = vmwarevif.get_vif_info(self._session, [ 640.630460] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 640.630846] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] for vif in network_info: [ 640.630846] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 640.630846] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] return self._sync_wrapper(fn, *args, **kwargs) [ 640.630846] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 640.630846] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] self.wait() [ 640.630846] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 640.630846] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] self[:] = self._gt.wait() [ 640.630846] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 640.630846] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] return self._exit_event.wait() [ 640.630846] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 640.630846] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] result = hub.switch() [ 640.630846] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 640.630846] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] return self.greenlet.switch() [ 640.631224] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 640.631224] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] result = function(*args, **kwargs) [ 640.631224] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 640.631224] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] return func(*args, **kwargs) [ 640.631224] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 640.631224] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] raise e [ 640.631224] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 640.631224] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] nwinfo = self.network_api.allocate_for_instance( [ 640.631224] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 640.631224] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] created_port_ids = self._update_ports_for_instance( [ 640.631224] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 640.631224] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] with excutils.save_and_reraise_exception(): [ 640.631224] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 640.631699] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] self.force_reraise() [ 640.631699] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 640.631699] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] raise self.value [ 640.631699] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 640.631699] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] updated_port = self._update_port( [ 640.631699] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 640.631699] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] _ensure_no_port_binding_failure(port) [ 640.631699] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 640.631699] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] raise exception.PortBindingFailed(port_id=port['id']) [ 640.631699] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] nova.exception.PortBindingFailed: Binding failed for port 4f7cdd79-dbcd-4758-a853-c2f8f2591818, please check neutron logs for more information. [ 640.631699] env[62753]: ERROR nova.compute.manager [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] [ 640.632038] env[62753]: DEBUG nova.compute.utils [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Binding failed for port 4f7cdd79-dbcd-4758-a853-c2f8f2591818, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 640.636020] env[62753]: DEBUG nova.compute.manager [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Build of instance b4e2a5f2-7225-47fa-883b-d892c92e3949 was re-scheduled: Binding failed for port 4f7cdd79-dbcd-4758-a853-c2f8f2591818, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 640.636020] env[62753]: DEBUG nova.compute.manager [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 640.636020] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Acquiring lock "refresh_cache-b4e2a5f2-7225-47fa-883b-d892c92e3949" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 640.636020] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Acquired lock "refresh_cache-b4e2a5f2-7225-47fa-883b-d892c92e3949" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.636326] env[62753]: DEBUG nova.network.neutron [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 640.636326] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.706s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 640.684986] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Releasing lock "refresh_cache-32dae012-7743-4efb-b39b-820c9ed42cc5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 640.686242] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 640.686242] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 640.686242] env[62753]: DEBUG oslo_concurrency.lockutils [req-91c4312d-8fdc-45cd-b11c-d1853ec7c816 req-7d1dfd9b-9646-455c-92d3-0f70a4c478e6 service nova] Acquired lock "refresh_cache-32dae012-7743-4efb-b39b-820c9ed42cc5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 640.686242] env[62753]: DEBUG nova.network.neutron [req-91c4312d-8fdc-45cd-b11c-d1853ec7c816 req-7d1dfd9b-9646-455c-92d3-0f70a4c478e6 service nova] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Refreshing network info cache for port 75747bb4-bc24-4d44-812f-8f0a74badb68 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 640.687818] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e185099d-f066-4c0f-900c-17d20bbe81fe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.700107] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d22af868-5ef9-499e-bb68-e43e342c911d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 640.727021] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 32dae012-7743-4efb-b39b-820c9ed42cc5 could not be found. [ 640.727021] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 640.727021] env[62753]: INFO nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Took 0.04 seconds to destroy the instance on the hypervisor. [ 640.727021] env[62753]: DEBUG oslo.service.loopingcall [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 640.727568] env[62753]: DEBUG nova.compute.manager [-] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 640.727568] env[62753]: DEBUG nova.network.neutron [-] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 640.747200] env[62753]: DEBUG nova.network.neutron [-] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 640.826931] env[62753]: INFO nova.compute.manager [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] [instance: 3cd890d9-4f45-404d-ba3e-7104abcf2b7d] Took 1.05 seconds to deallocate network for instance. [ 640.830203] env[62753]: DEBUG nova.compute.manager [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 641.185994] env[62753]: DEBUG nova.network.neutron [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.209791] env[62753]: DEBUG nova.network.neutron [req-91c4312d-8fdc-45cd-b11c-d1853ec7c816 req-7d1dfd9b-9646-455c-92d3-0f70a4c478e6 service nova] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.249816] env[62753]: DEBUG nova.network.neutron [-] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.306628] env[62753]: DEBUG nova.network.neutron [req-91c4312d-8fdc-45cd-b11c-d1853ec7c816 req-7d1dfd9b-9646-455c-92d3-0f70a4c478e6 service nova] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.337405] env[62753]: DEBUG nova.network.neutron [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 641.359268] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.574502] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b5a40d7-542a-4d36-9886-521301a7c94f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.585052] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e438bfc-d828-47cc-b3c2-ba98ba312dd1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.614719] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-685cea8d-8006-48dc-898a-ec40824af508 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.622098] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26ad870e-d104-4fd1-8e92-ba0a09e62b41 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 641.635974] env[62753]: DEBUG nova.compute.provider_tree [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 641.752438] env[62753]: INFO nova.compute.manager [-] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Took 1.02 seconds to deallocate network for instance. [ 641.758870] env[62753]: DEBUG nova.compute.claims [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 641.758870] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.814125] env[62753]: DEBUG oslo_concurrency.lockutils [req-91c4312d-8fdc-45cd-b11c-d1853ec7c816 req-7d1dfd9b-9646-455c-92d3-0f70a4c478e6 service nova] Releasing lock "refresh_cache-32dae012-7743-4efb-b39b-820c9ed42cc5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.817826] env[62753]: DEBUG oslo_concurrency.lockutils [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Acquiring lock "b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 641.818057] env[62753]: DEBUG oslo_concurrency.lockutils [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Lock "b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 641.846699] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Releasing lock "refresh_cache-b4e2a5f2-7225-47fa-883b-d892c92e3949" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 641.846929] env[62753]: DEBUG nova.compute.manager [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 641.847099] env[62753]: DEBUG nova.compute.manager [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 641.847264] env[62753]: DEBUG nova.network.neutron [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 641.873207] env[62753]: INFO nova.scheduler.client.report [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Deleted allocations for instance 3cd890d9-4f45-404d-ba3e-7104abcf2b7d [ 641.886762] env[62753]: DEBUG nova.network.neutron [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 641.898161] env[62753]: DEBUG nova.compute.manager [req-89346b80-6b33-415f-a0ca-981a77a3cc9c req-af13db6d-a04e-4b0b-b1e2-0d3e8a5c8db4 service nova] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Received event network-vif-deleted-75747bb4-bc24-4d44-812f-8f0a74badb68 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 642.141097] env[62753]: DEBUG nova.scheduler.client.report [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 642.387013] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9de9702-8cc5-41f0-8b50-8394423f3dea tempest-VolumesAssistedSnapshotsTest-2082091531 tempest-VolumesAssistedSnapshotsTest-2082091531-project-member] Lock "3cd890d9-4f45-404d-ba3e-7104abcf2b7d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 51.494s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.388546] env[62753]: DEBUG nova.network.neutron [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 642.650211] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.013s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 642.650211] env[62753]: ERROR nova.compute.manager [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 99c15a1c-c10e-4c95-bf5f-695f7631aef8, please check neutron logs for more information. [ 642.650211] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Traceback (most recent call last): [ 642.650211] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 642.650211] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] self.driver.spawn(context, instance, image_meta, [ 642.650211] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 642.650211] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] self._vmops.spawn(context, instance, image_meta, injected_files, [ 642.650211] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 642.650211] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] vm_ref = self.build_virtual_machine(instance, [ 642.650599] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 642.650599] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] vif_infos = vmwarevif.get_vif_info(self._session, [ 642.650599] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 642.650599] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] for vif in network_info: [ 642.650599] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 642.650599] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] return self._sync_wrapper(fn, *args, **kwargs) [ 642.650599] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 642.650599] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] self.wait() [ 642.650599] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 642.650599] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] self[:] = self._gt.wait() [ 642.650599] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 642.650599] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] return self._exit_event.wait() [ 642.650599] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 642.651078] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] result = hub.switch() [ 642.651078] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 642.651078] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] return self.greenlet.switch() [ 642.651078] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 642.651078] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] result = function(*args, **kwargs) [ 642.651078] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 642.651078] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] return func(*args, **kwargs) [ 642.651078] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 642.651078] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] raise e [ 642.651078] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 642.651078] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] nwinfo = self.network_api.allocate_for_instance( [ 642.651078] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 642.651078] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] created_port_ids = self._update_ports_for_instance( [ 642.651463] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 642.651463] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] with excutils.save_and_reraise_exception(): [ 642.651463] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 642.651463] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] self.force_reraise() [ 642.651463] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 642.651463] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] raise self.value [ 642.651463] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 642.651463] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] updated_port = self._update_port( [ 642.651463] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 642.651463] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] _ensure_no_port_binding_failure(port) [ 642.651463] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 642.651463] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] raise exception.PortBindingFailed(port_id=port['id']) [ 642.651869] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] nova.exception.PortBindingFailed: Binding failed for port 99c15a1c-c10e-4c95-bf5f-695f7631aef8, please check neutron logs for more information. [ 642.651869] env[62753]: ERROR nova.compute.manager [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] [ 642.651869] env[62753]: DEBUG nova.compute.utils [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Binding failed for port 99c15a1c-c10e-4c95-bf5f-695f7631aef8, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 642.651979] env[62753]: DEBUG oslo_concurrency.lockutils [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.580s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 642.653834] env[62753]: INFO nova.compute.claims [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 642.656629] env[62753]: DEBUG nova.compute.manager [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Build of instance e5faf8e5-a5f6-4484-9304-7bbef03b33fc was re-scheduled: Binding failed for port 99c15a1c-c10e-4c95-bf5f-695f7631aef8, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 642.657321] env[62753]: DEBUG nova.compute.manager [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 642.657321] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Acquiring lock "refresh_cache-e5faf8e5-a5f6-4484-9304-7bbef03b33fc" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 642.657459] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Acquired lock "refresh_cache-e5faf8e5-a5f6-4484-9304-7bbef03b33fc" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 642.658171] env[62753]: DEBUG nova.network.neutron [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 642.888407] env[62753]: DEBUG nova.compute.manager [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 642.895118] env[62753]: INFO nova.compute.manager [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: b4e2a5f2-7225-47fa-883b-d892c92e3949] Took 1.05 seconds to deallocate network for instance. [ 643.192133] env[62753]: DEBUG nova.network.neutron [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.321944] env[62753]: DEBUG nova.network.neutron [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 643.424326] env[62753]: DEBUG oslo_concurrency.lockutils [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 643.826849] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Releasing lock "refresh_cache-e5faf8e5-a5f6-4484-9304-7bbef03b33fc" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 643.827519] env[62753]: DEBUG nova.compute.manager [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 643.827519] env[62753]: DEBUG nova.compute.manager [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 643.827519] env[62753]: DEBUG nova.network.neutron [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 643.852873] env[62753]: DEBUG nova.network.neutron [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 643.955599] env[62753]: INFO nova.scheduler.client.report [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Deleted allocations for instance b4e2a5f2-7225-47fa-883b-d892c92e3949 [ 644.204124] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-942b8ed8-d929-45f6-aea7-4abe4b000703 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.212043] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aff3bd4-7fea-44b2-8892-e79006d58990 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.251214] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e6b0007-6317-4e1d-b88f-80182aff49d1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.263852] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-795a5011-47f8-4d82-b570-e1c4c7085d05 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 644.278370] env[62753]: DEBUG nova.compute.provider_tree [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 644.357813] env[62753]: DEBUG nova.network.neutron [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 644.467461] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ec605ece-e661-4416-9826-4df92a2aecff tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Lock "b4e2a5f2-7225-47fa-883b-d892c92e3949" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.063s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 644.782626] env[62753]: DEBUG nova.scheduler.client.report [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 644.866920] env[62753]: INFO nova.compute.manager [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] [instance: e5faf8e5-a5f6-4484-9304-7bbef03b33fc] Took 1.04 seconds to deallocate network for instance. [ 644.969956] env[62753]: DEBUG nova.compute.manager [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 645.288440] env[62753]: DEBUG oslo_concurrency.lockutils [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.636s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 645.288984] env[62753]: DEBUG nova.compute.manager [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 645.291680] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.127s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.293064] env[62753]: INFO nova.compute.claims [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 645.462241] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "358d1c98-28ee-4f18-bd42-89a1e7687d03" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.463532] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "358d1c98-28ee-4f18-bd42-89a1e7687d03" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 645.491403] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 645.797147] env[62753]: DEBUG nova.compute.utils [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 645.805033] env[62753]: DEBUG nova.compute.manager [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 645.805033] env[62753]: DEBUG nova.network.neutron [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 645.904543] env[62753]: INFO nova.scheduler.client.report [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Deleted allocations for instance e5faf8e5-a5f6-4484-9304-7bbef03b33fc [ 645.945872] env[62753]: DEBUG nova.policy [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ca7ddbd90a454b5ab41ebae1a1f4f33f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '53e75036fc0d4d689316e9e4fde34731', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 646.303058] env[62753]: DEBUG nova.compute.manager [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 646.416953] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0e9a7252-d8e5-4a49-9712-a89d6a25ca26 tempest-ListImageFiltersTestJSON-1920704467 tempest-ListImageFiltersTestJSON-1920704467-project-member] Lock "e5faf8e5-a5f6-4484-9304-7bbef03b33fc" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.667s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 646.733080] env[62753]: DEBUG nova.network.neutron [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Successfully created port: e72e4a03-e91c-4040-92ce-0c6ed825b671 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 646.745684] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8362feb-162d-415c-9a7c-646d6acac9f3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.752869] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-468febdf-6458-47f2-95e6-798b8b81fd06 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.798969] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b58a8f20-f6ac-4f16-8613-d0ccd4b7814c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.808583] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-131e2d7f-4b08-4bf6-88f4-7c78ed0bcd78 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 646.826236] env[62753]: DEBUG nova.compute.provider_tree [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 646.861582] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "365fe802-f5e1-47d6-807e-a3b1027def02" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 646.862228] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "365fe802-f5e1-47d6-807e-a3b1027def02" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 646.926543] env[62753]: DEBUG nova.compute.manager [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 647.321020] env[62753]: DEBUG nova.compute.manager [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 647.332176] env[62753]: DEBUG nova.scheduler.client.report [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 647.356582] env[62753]: DEBUG nova.virt.hardware [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 647.356582] env[62753]: DEBUG nova.virt.hardware [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 647.356582] env[62753]: DEBUG nova.virt.hardware [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 647.356814] env[62753]: DEBUG nova.virt.hardware [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 647.356814] env[62753]: DEBUG nova.virt.hardware [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 647.356814] env[62753]: DEBUG nova.virt.hardware [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 647.356814] env[62753]: DEBUG nova.virt.hardware [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 647.357409] env[62753]: DEBUG nova.virt.hardware [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 647.358367] env[62753]: DEBUG nova.virt.hardware [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 647.358367] env[62753]: DEBUG nova.virt.hardware [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 647.358367] env[62753]: DEBUG nova.virt.hardware [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 647.359530] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9204822a-3e9c-4cf3-ac1c-9fd3a3ed397a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.372020] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-824db4b6-a4b6-4430-bfbb-f0d0c9d5f80c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 647.452247] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 647.835872] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.544s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 647.836495] env[62753]: DEBUG nova.compute.manager [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 647.843678] env[62753]: DEBUG oslo_concurrency.lockutils [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.111s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 647.848783] env[62753]: INFO nova.compute.claims [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 648.347844] env[62753]: DEBUG nova.compute.utils [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 648.347844] env[62753]: DEBUG nova.compute.manager [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 648.347844] env[62753]: DEBUG nova.network.neutron [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 648.423564] env[62753]: DEBUG nova.policy [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'c8c9bf3ec728415eb592a246d7c30e77', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '25b07d71a5bd443296fedf91dfbd020e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 648.701405] env[62753]: DEBUG nova.compute.manager [req-09220026-b1af-4fc2-be89-fe32f2a9d63d req-89868172-581c-4bab-814b-e62d8dfb0603 service nova] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Received event network-changed-e72e4a03-e91c-4040-92ce-0c6ed825b671 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 648.701724] env[62753]: DEBUG nova.compute.manager [req-09220026-b1af-4fc2-be89-fe32f2a9d63d req-89868172-581c-4bab-814b-e62d8dfb0603 service nova] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Refreshing instance network info cache due to event network-changed-e72e4a03-e91c-4040-92ce-0c6ed825b671. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 648.702034] env[62753]: DEBUG oslo_concurrency.lockutils [req-09220026-b1af-4fc2-be89-fe32f2a9d63d req-89868172-581c-4bab-814b-e62d8dfb0603 service nova] Acquiring lock "refresh_cache-2d028151-3d84-4195-9f93-28287dbfda09" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 648.702155] env[62753]: DEBUG oslo_concurrency.lockutils [req-09220026-b1af-4fc2-be89-fe32f2a9d63d req-89868172-581c-4bab-814b-e62d8dfb0603 service nova] Acquired lock "refresh_cache-2d028151-3d84-4195-9f93-28287dbfda09" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 648.702373] env[62753]: DEBUG nova.network.neutron [req-09220026-b1af-4fc2-be89-fe32f2a9d63d req-89868172-581c-4bab-814b-e62d8dfb0603 service nova] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Refreshing network info cache for port e72e4a03-e91c-4040-92ce-0c6ed825b671 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 648.854225] env[62753]: DEBUG nova.compute.manager [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 648.958094] env[62753]: DEBUG nova.network.neutron [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Successfully created port: 05ad2098-191c-4e2e-a2de-94432f0b7a25 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 649.046027] env[62753]: ERROR nova.compute.manager [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e72e4a03-e91c-4040-92ce-0c6ed825b671, please check neutron logs for more information. [ 649.046027] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 649.046027] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.046027] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 649.046027] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.046027] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 649.046027] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.046027] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 649.046027] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.046027] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 649.046027] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.046027] env[62753]: ERROR nova.compute.manager raise self.value [ 649.046027] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.046027] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 649.046027] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.046027] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 649.046553] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.046553] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 649.046553] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e72e4a03-e91c-4040-92ce-0c6ed825b671, please check neutron logs for more information. [ 649.046553] env[62753]: ERROR nova.compute.manager [ 649.046553] env[62753]: Traceback (most recent call last): [ 649.046553] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 649.046553] env[62753]: listener.cb(fileno) [ 649.046553] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.046553] env[62753]: result = function(*args, **kwargs) [ 649.046553] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 649.046553] env[62753]: return func(*args, **kwargs) [ 649.046553] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 649.046553] env[62753]: raise e [ 649.046553] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.046553] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 649.046553] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.046553] env[62753]: created_port_ids = self._update_ports_for_instance( [ 649.046553] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.046553] env[62753]: with excutils.save_and_reraise_exception(): [ 649.046553] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.046553] env[62753]: self.force_reraise() [ 649.046553] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.046553] env[62753]: raise self.value [ 649.046553] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.046553] env[62753]: updated_port = self._update_port( [ 649.046553] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.046553] env[62753]: _ensure_no_port_binding_failure(port) [ 649.046553] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.046553] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 649.047172] env[62753]: nova.exception.PortBindingFailed: Binding failed for port e72e4a03-e91c-4040-92ce-0c6ed825b671, please check neutron logs for more information. [ 649.047172] env[62753]: Removing descriptor: 14 [ 649.047172] env[62753]: ERROR nova.compute.manager [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e72e4a03-e91c-4040-92ce-0c6ed825b671, please check neutron logs for more information. [ 649.047172] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Traceback (most recent call last): [ 649.047172] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 649.047172] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] yield resources [ 649.047172] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 649.047172] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] self.driver.spawn(context, instance, image_meta, [ 649.047172] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 649.047172] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] self._vmops.spawn(context, instance, image_meta, injected_files, [ 649.047172] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 649.047172] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] vm_ref = self.build_virtual_machine(instance, [ 649.047425] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 649.047425] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] vif_infos = vmwarevif.get_vif_info(self._session, [ 649.047425] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 649.047425] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] for vif in network_info: [ 649.047425] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 649.047425] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] return self._sync_wrapper(fn, *args, **kwargs) [ 649.047425] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 649.047425] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] self.wait() [ 649.047425] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 649.047425] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] self[:] = self._gt.wait() [ 649.047425] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 649.047425] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] return self._exit_event.wait() [ 649.047425] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 649.047697] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] result = hub.switch() [ 649.047697] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 649.047697] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] return self.greenlet.switch() [ 649.047697] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 649.047697] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] result = function(*args, **kwargs) [ 649.047697] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 649.047697] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] return func(*args, **kwargs) [ 649.047697] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 649.047697] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] raise e [ 649.047697] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 649.047697] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] nwinfo = self.network_api.allocate_for_instance( [ 649.047697] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 649.047697] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] created_port_ids = self._update_ports_for_instance( [ 649.047974] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 649.047974] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] with excutils.save_and_reraise_exception(): [ 649.047974] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 649.047974] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] self.force_reraise() [ 649.047974] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 649.047974] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] raise self.value [ 649.047974] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 649.047974] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] updated_port = self._update_port( [ 649.047974] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 649.047974] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] _ensure_no_port_binding_failure(port) [ 649.047974] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 649.047974] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] raise exception.PortBindingFailed(port_id=port['id']) [ 649.048228] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] nova.exception.PortBindingFailed: Binding failed for port e72e4a03-e91c-4040-92ce-0c6ed825b671, please check neutron logs for more information. [ 649.048228] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] [ 649.048228] env[62753]: INFO nova.compute.manager [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Terminating instance [ 649.048228] env[62753]: DEBUG oslo_concurrency.lockutils [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Acquiring lock "refresh_cache-2d028151-3d84-4195-9f93-28287dbfda09" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 649.242598] env[62753]: DEBUG nova.network.neutron [req-09220026-b1af-4fc2-be89-fe32f2a9d63d req-89868172-581c-4bab-814b-e62d8dfb0603 service nova] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 649.336151] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7818b281-fcf4-4496-b60a-6c6bbe431a85 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.349555] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd77abd9-6590-4de3-b0ad-b618e4b90bf5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.396570] env[62753]: DEBUG nova.network.neutron [req-09220026-b1af-4fc2-be89-fe32f2a9d63d req-89868172-581c-4bab-814b-e62d8dfb0603 service nova] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 649.398068] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ccef0f4-34ed-43d1-88aa-9178771c37b2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.405864] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-535599fc-75b9-4f19-ba9f-8c332f12a606 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.421843] env[62753]: DEBUG nova.compute.provider_tree [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 649.897368] env[62753]: DEBUG nova.compute.manager [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 649.901960] env[62753]: DEBUG oslo_concurrency.lockutils [req-09220026-b1af-4fc2-be89-fe32f2a9d63d req-89868172-581c-4bab-814b-e62d8dfb0603 service nova] Releasing lock "refresh_cache-2d028151-3d84-4195-9f93-28287dbfda09" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 649.902369] env[62753]: DEBUG oslo_concurrency.lockutils [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Acquired lock "refresh_cache-2d028151-3d84-4195-9f93-28287dbfda09" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 649.902547] env[62753]: DEBUG nova.network.neutron [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 649.924540] env[62753]: DEBUG nova.scheduler.client.report [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 649.937480] env[62753]: DEBUG nova.virt.hardware [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:26:53Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='1365876506',id=24,is_public=True,memory_mb=192,name='tempest-flavor_with_ephemeral_0-684804906',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 649.937825] env[62753]: DEBUG nova.virt.hardware [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 649.937908] env[62753]: DEBUG nova.virt.hardware [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 649.938207] env[62753]: DEBUG nova.virt.hardware [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 649.938207] env[62753]: DEBUG nova.virt.hardware [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 649.938352] env[62753]: DEBUG nova.virt.hardware [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 649.938586] env[62753]: DEBUG nova.virt.hardware [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 649.938748] env[62753]: DEBUG nova.virt.hardware [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 649.938917] env[62753]: DEBUG nova.virt.hardware [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 649.939138] env[62753]: DEBUG nova.virt.hardware [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 649.939268] env[62753]: DEBUG nova.virt.hardware [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 649.941303] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7764e856-18b1-4097-a361-189604db70ba {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 649.951318] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff5a2afd-1a1a-4c11-80d0-d012f4db8bf2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 650.423266] env[62753]: DEBUG nova.compute.manager [req-4ef1dae9-68d1-47f4-acec-aa8e859d2885 req-39d22c6b-686e-4735-bff3-1f10af294632 service nova] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Received event network-changed-05ad2098-191c-4e2e-a2de-94432f0b7a25 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 650.423698] env[62753]: DEBUG nova.compute.manager [req-4ef1dae9-68d1-47f4-acec-aa8e859d2885 req-39d22c6b-686e-4735-bff3-1f10af294632 service nova] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Refreshing instance network info cache due to event network-changed-05ad2098-191c-4e2e-a2de-94432f0b7a25. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 650.424014] env[62753]: DEBUG oslo_concurrency.lockutils [req-4ef1dae9-68d1-47f4-acec-aa8e859d2885 req-39d22c6b-686e-4735-bff3-1f10af294632 service nova] Acquiring lock "refresh_cache-989db77e-169f-4b3b-a0b9-c4417f4e89c4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.424281] env[62753]: DEBUG oslo_concurrency.lockutils [req-4ef1dae9-68d1-47f4-acec-aa8e859d2885 req-39d22c6b-686e-4735-bff3-1f10af294632 service nova] Acquired lock "refresh_cache-989db77e-169f-4b3b-a0b9-c4417f4e89c4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 650.424464] env[62753]: DEBUG nova.network.neutron [req-4ef1dae9-68d1-47f4-acec-aa8e859d2885 req-39d22c6b-686e-4735-bff3-1f10af294632 service nova] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Refreshing network info cache for port 05ad2098-191c-4e2e-a2de-94432f0b7a25 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 650.430888] env[62753]: DEBUG oslo_concurrency.lockutils [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.587s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 650.431436] env[62753]: DEBUG nova.compute.manager [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 650.434144] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.175s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 650.453520] env[62753]: DEBUG nova.network.neutron [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 650.571126] env[62753]: ERROR nova.compute.manager [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 05ad2098-191c-4e2e-a2de-94432f0b7a25, please check neutron logs for more information. [ 650.571126] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 650.571126] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.571126] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 650.571126] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 650.571126] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 650.571126] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 650.571126] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 650.571126] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.571126] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 650.571126] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.571126] env[62753]: ERROR nova.compute.manager raise self.value [ 650.571126] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 650.571126] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 650.571126] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.571126] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 650.571629] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.571629] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 650.571629] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 05ad2098-191c-4e2e-a2de-94432f0b7a25, please check neutron logs for more information. [ 650.571629] env[62753]: ERROR nova.compute.manager [ 650.571629] env[62753]: Traceback (most recent call last): [ 650.571629] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 650.571629] env[62753]: listener.cb(fileno) [ 650.571629] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.571629] env[62753]: result = function(*args, **kwargs) [ 650.571629] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 650.571629] env[62753]: return func(*args, **kwargs) [ 650.571629] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.571629] env[62753]: raise e [ 650.571629] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.571629] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 650.571629] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 650.571629] env[62753]: created_port_ids = self._update_ports_for_instance( [ 650.571629] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 650.571629] env[62753]: with excutils.save_and_reraise_exception(): [ 650.571629] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.571629] env[62753]: self.force_reraise() [ 650.571629] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.571629] env[62753]: raise self.value [ 650.571629] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 650.571629] env[62753]: updated_port = self._update_port( [ 650.571629] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.571629] env[62753]: _ensure_no_port_binding_failure(port) [ 650.571629] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.571629] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 650.572266] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 05ad2098-191c-4e2e-a2de-94432f0b7a25, please check neutron logs for more information. [ 650.572266] env[62753]: Removing descriptor: 16 [ 650.572266] env[62753]: ERROR nova.compute.manager [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 05ad2098-191c-4e2e-a2de-94432f0b7a25, please check neutron logs for more information. [ 650.572266] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Traceback (most recent call last): [ 650.572266] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 650.572266] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] yield resources [ 650.572266] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 650.572266] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] self.driver.spawn(context, instance, image_meta, [ 650.572266] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 650.572266] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 650.572266] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 650.572266] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] vm_ref = self.build_virtual_machine(instance, [ 650.572545] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 650.572545] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] vif_infos = vmwarevif.get_vif_info(self._session, [ 650.572545] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 650.572545] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] for vif in network_info: [ 650.572545] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 650.572545] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] return self._sync_wrapper(fn, *args, **kwargs) [ 650.572545] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 650.572545] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] self.wait() [ 650.572545] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 650.572545] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] self[:] = self._gt.wait() [ 650.572545] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 650.572545] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] return self._exit_event.wait() [ 650.572545] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 650.572841] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] result = hub.switch() [ 650.572841] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 650.572841] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] return self.greenlet.switch() [ 650.572841] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 650.572841] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] result = function(*args, **kwargs) [ 650.572841] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 650.572841] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] return func(*args, **kwargs) [ 650.572841] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 650.572841] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] raise e [ 650.572841] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 650.572841] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] nwinfo = self.network_api.allocate_for_instance( [ 650.572841] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 650.572841] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] created_port_ids = self._update_ports_for_instance( [ 650.573160] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 650.573160] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] with excutils.save_and_reraise_exception(): [ 650.573160] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 650.573160] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] self.force_reraise() [ 650.573160] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 650.573160] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] raise self.value [ 650.573160] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 650.573160] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] updated_port = self._update_port( [ 650.573160] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 650.573160] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] _ensure_no_port_binding_failure(port) [ 650.573160] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 650.573160] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] raise exception.PortBindingFailed(port_id=port['id']) [ 650.573424] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] nova.exception.PortBindingFailed: Binding failed for port 05ad2098-191c-4e2e-a2de-94432f0b7a25, please check neutron logs for more information. [ 650.573424] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] [ 650.573424] env[62753]: INFO nova.compute.manager [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Terminating instance [ 650.574855] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Acquiring lock "refresh_cache-989db77e-169f-4b3b-a0b9-c4417f4e89c4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 650.627209] env[62753]: DEBUG nova.network.neutron [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 650.942432] env[62753]: DEBUG nova.compute.utils [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 650.944462] env[62753]: DEBUG nova.compute.manager [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 650.944648] env[62753]: DEBUG nova.network.neutron [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 651.017603] env[62753]: DEBUG nova.policy [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3e670a21b97044bba2b7590c56129cf2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3451ff7d7a8e4dbea0c7c5aef6388a35', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 651.051451] env[62753]: DEBUG nova.network.neutron [req-4ef1dae9-68d1-47f4-acec-aa8e859d2885 req-39d22c6b-686e-4735-bff3-1f10af294632 service nova] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.080188] env[62753]: DEBUG nova.compute.manager [req-da028380-8f42-4faa-989a-9945cef77214 req-ecb012fe-2bdc-4648-965c-8592c884d4c5 service nova] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Received event network-vif-deleted-e72e4a03-e91c-4040-92ce-0c6ed825b671 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 651.130099] env[62753]: DEBUG oslo_concurrency.lockutils [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Releasing lock "refresh_cache-2d028151-3d84-4195-9f93-28287dbfda09" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.130568] env[62753]: DEBUG nova.compute.manager [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 651.131328] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 651.131328] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-06ad90ba-6ec9-497e-aec5-e131c3c017d1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.141060] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3181a043-7ddd-4532-a96a-a41b2a495aec {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.167298] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 2d028151-3d84-4195-9f93-28287dbfda09 could not be found. [ 651.167513] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 651.167691] env[62753]: INFO nova.compute.manager [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Took 0.04 seconds to destroy the instance on the hypervisor. [ 651.167933] env[62753]: DEBUG oslo.service.loopingcall [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 651.168154] env[62753]: DEBUG nova.compute.manager [-] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 651.168238] env[62753]: DEBUG nova.network.neutron [-] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 651.188829] env[62753]: DEBUG nova.network.neutron [-] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 651.233289] env[62753]: DEBUG nova.network.neutron [req-4ef1dae9-68d1-47f4-acec-aa8e859d2885 req-39d22c6b-686e-4735-bff3-1f10af294632 service nova] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.369729] env[62753]: DEBUG nova.network.neutron [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Successfully created port: 43989f12-9ac8-484f-a0e9-ad981ecf1fca {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 651.409301] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d3e2cb-cd10-4678-b01b-e6bf39834a39 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.419147] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7722a746-21f2-4423-ad57-eae09f9af674 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.451315] env[62753]: DEBUG nova.compute.manager [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 651.455023] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7081f86c-5284-4311-9af4-5f2d8bdc0eac {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.464751] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0e6dda-452b-4bc3-a555-c58e162054dc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 651.478019] env[62753]: DEBUG nova.compute.provider_tree [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 651.691335] env[62753]: DEBUG nova.network.neutron [-] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 651.738428] env[62753]: DEBUG oslo_concurrency.lockutils [req-4ef1dae9-68d1-47f4-acec-aa8e859d2885 req-39d22c6b-686e-4735-bff3-1f10af294632 service nova] Releasing lock "refresh_cache-989db77e-169f-4b3b-a0b9-c4417f4e89c4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 651.739633] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Acquired lock "refresh_cache-989db77e-169f-4b3b-a0b9-c4417f4e89c4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 651.739833] env[62753]: DEBUG nova.network.neutron [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 651.914051] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Acquiring lock "8664ed5c-ecb9-4795-8499-31198cfd0450" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 651.914393] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Lock "8664ed5c-ecb9-4795-8499-31198cfd0450" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 651.981500] env[62753]: DEBUG nova.scheduler.client.report [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 652.195162] env[62753]: INFO nova.compute.manager [-] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Took 1.03 seconds to deallocate network for instance. [ 652.199321] env[62753]: DEBUG nova.compute.claims [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 652.199321] env[62753]: DEBUG oslo_concurrency.lockutils [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.269926] env[62753]: DEBUG nova.network.neutron [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 652.324018] env[62753]: ERROR nova.compute.manager [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 43989f12-9ac8-484f-a0e9-ad981ecf1fca, please check neutron logs for more information. [ 652.324018] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 652.324018] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.324018] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 652.324018] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 652.324018] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 652.324018] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 652.324018] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 652.324018] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.324018] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 652.324018] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.324018] env[62753]: ERROR nova.compute.manager raise self.value [ 652.324018] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 652.324018] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 652.324018] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.324018] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 652.324468] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.324468] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 652.324468] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 43989f12-9ac8-484f-a0e9-ad981ecf1fca, please check neutron logs for more information. [ 652.324468] env[62753]: ERROR nova.compute.manager [ 652.324468] env[62753]: Traceback (most recent call last): [ 652.324468] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 652.324468] env[62753]: listener.cb(fileno) [ 652.324468] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.324468] env[62753]: result = function(*args, **kwargs) [ 652.324468] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 652.324468] env[62753]: return func(*args, **kwargs) [ 652.324468] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.324468] env[62753]: raise e [ 652.324468] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.324468] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 652.324468] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 652.324468] env[62753]: created_port_ids = self._update_ports_for_instance( [ 652.324468] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 652.324468] env[62753]: with excutils.save_and_reraise_exception(): [ 652.324468] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.324468] env[62753]: self.force_reraise() [ 652.324468] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.324468] env[62753]: raise self.value [ 652.324468] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 652.324468] env[62753]: updated_port = self._update_port( [ 652.324468] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.324468] env[62753]: _ensure_no_port_binding_failure(port) [ 652.324468] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.324468] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 652.329397] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 43989f12-9ac8-484f-a0e9-ad981ecf1fca, please check neutron logs for more information. [ 652.329397] env[62753]: Removing descriptor: 16 [ 652.465332] env[62753]: DEBUG nova.compute.manager [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 652.479311] env[62753]: DEBUG nova.network.neutron [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 652.490099] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.054s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 652.490099] env[62753]: ERROR nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8f485a5c-58db-45f9-8fc7-928e5e7d8eb9, please check neutron logs for more information. [ 652.490099] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Traceback (most recent call last): [ 652.490099] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 652.490099] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] self.driver.spawn(context, instance, image_meta, [ 652.490099] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 652.490099] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.490099] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.490099] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] vm_ref = self.build_virtual_machine(instance, [ 652.490385] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.490385] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.490385] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.490385] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] for vif in network_info: [ 652.490385] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 652.490385] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] return self._sync_wrapper(fn, *args, **kwargs) [ 652.490385] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 652.490385] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] self.wait() [ 652.490385] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 652.490385] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] self[:] = self._gt.wait() [ 652.490385] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.490385] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] return self._exit_event.wait() [ 652.490385] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 652.490701] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] result = hub.switch() [ 652.490701] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 652.490701] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] return self.greenlet.switch() [ 652.490701] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.490701] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] result = function(*args, **kwargs) [ 652.490701] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 652.490701] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] return func(*args, **kwargs) [ 652.490701] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.490701] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] raise e [ 652.490701] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.490701] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] nwinfo = self.network_api.allocate_for_instance( [ 652.490701] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 652.490701] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] created_port_ids = self._update_ports_for_instance( [ 652.490961] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 652.490961] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] with excutils.save_and_reraise_exception(): [ 652.490961] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.490961] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] self.force_reraise() [ 652.490961] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.490961] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] raise self.value [ 652.490961] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 652.490961] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] updated_port = self._update_port( [ 652.490961] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.490961] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] _ensure_no_port_binding_failure(port) [ 652.490961] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.490961] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] raise exception.PortBindingFailed(port_id=port['id']) [ 652.491225] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] nova.exception.PortBindingFailed: Binding failed for port 8f485a5c-58db-45f9-8fc7-928e5e7d8eb9, please check neutron logs for more information. [ 652.491225] env[62753]: ERROR nova.compute.manager [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] [ 652.491225] env[62753]: DEBUG nova.compute.utils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Binding failed for port 8f485a5c-58db-45f9-8fc7-928e5e7d8eb9, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 652.491225] env[62753]: DEBUG oslo_concurrency.lockutils [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.570s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.492838] env[62753]: INFO nova.compute.claims [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 652.495519] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Build of instance 11cf82b0-ea78-40b2-b737-dea8b8703ac3 was re-scheduled: Binding failed for port 8f485a5c-58db-45f9-8fc7-928e5e7d8eb9, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 652.496059] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 652.496281] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "refresh_cache-11cf82b0-ea78-40b2-b737-dea8b8703ac3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.496438] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquired lock "refresh_cache-11cf82b0-ea78-40b2-b737-dea8b8703ac3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.496667] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 652.512984] env[62753]: DEBUG nova.virt.hardware [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 652.513727] env[62753]: DEBUG nova.virt.hardware [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 652.513727] env[62753]: DEBUG nova.virt.hardware [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 652.513856] env[62753]: DEBUG nova.virt.hardware [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 652.513967] env[62753]: DEBUG nova.virt.hardware [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 652.514131] env[62753]: DEBUG nova.virt.hardware [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 652.514366] env[62753]: DEBUG nova.virt.hardware [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 652.514959] env[62753]: DEBUG nova.virt.hardware [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 652.514959] env[62753]: DEBUG nova.virt.hardware [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 652.514959] env[62753]: DEBUG nova.virt.hardware [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 652.514959] env[62753]: DEBUG nova.virt.hardware [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 652.516133] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8c25c8e-8219-4239-893b-a59d08f53ba5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.525633] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83468b41-d6f4-496c-a4f3-80a3392ae54b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.541967] env[62753]: ERROR nova.compute.manager [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 43989f12-9ac8-484f-a0e9-ad981ecf1fca, please check neutron logs for more information. [ 652.541967] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Traceback (most recent call last): [ 652.541967] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 652.541967] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] yield resources [ 652.541967] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 652.541967] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] self.driver.spawn(context, instance, image_meta, [ 652.541967] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 652.541967] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] self._vmops.spawn(context, instance, image_meta, injected_files, [ 652.541967] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 652.541967] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] vm_ref = self.build_virtual_machine(instance, [ 652.541967] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 652.542294] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] vif_infos = vmwarevif.get_vif_info(self._session, [ 652.542294] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 652.542294] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] for vif in network_info: [ 652.542294] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 652.542294] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] return self._sync_wrapper(fn, *args, **kwargs) [ 652.542294] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 652.542294] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] self.wait() [ 652.542294] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 652.542294] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] self[:] = self._gt.wait() [ 652.542294] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 652.542294] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] return self._exit_event.wait() [ 652.542294] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 652.542294] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] current.throw(*self._exc) [ 652.542612] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 652.542612] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] result = function(*args, **kwargs) [ 652.542612] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 652.542612] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] return func(*args, **kwargs) [ 652.542612] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 652.542612] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] raise e [ 652.542612] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 652.542612] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] nwinfo = self.network_api.allocate_for_instance( [ 652.542612] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 652.542612] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] created_port_ids = self._update_ports_for_instance( [ 652.542612] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 652.542612] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] with excutils.save_and_reraise_exception(): [ 652.542612] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 652.542894] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] self.force_reraise() [ 652.542894] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 652.542894] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] raise self.value [ 652.542894] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 652.542894] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] updated_port = self._update_port( [ 652.542894] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 652.542894] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] _ensure_no_port_binding_failure(port) [ 652.542894] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 652.542894] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] raise exception.PortBindingFailed(port_id=port['id']) [ 652.542894] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] nova.exception.PortBindingFailed: Binding failed for port 43989f12-9ac8-484f-a0e9-ad981ecf1fca, please check neutron logs for more information. [ 652.542894] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] [ 652.542894] env[62753]: INFO nova.compute.manager [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Terminating instance [ 652.543236] env[62753]: DEBUG oslo_concurrency.lockutils [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Acquiring lock "refresh_cache-3c8e531e-033c-4857-9865-b1024e143d44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.543278] env[62753]: DEBUG oslo_concurrency.lockutils [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Acquired lock "refresh_cache-3c8e531e-033c-4857-9865-b1024e143d44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 652.543438] env[62753]: DEBUG nova.network.neutron [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 652.624841] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Acquiring lock "73f7840f-6e3e-4a4e-b7f5-bffe3009aae6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 652.624841] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Lock "73f7840f-6e3e-4a4e-b7f5-bffe3009aae6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 652.626820] env[62753]: DEBUG nova.compute.manager [req-8b9bacd1-ddfc-4063-ab78-2d9ce3190283 req-546e745e-6bf9-4f25-ab72-983dbc874b50 service nova] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Received event network-vif-deleted-05ad2098-191c-4e2e-a2de-94432f0b7a25 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 652.627311] env[62753]: DEBUG nova.compute.manager [req-8b9bacd1-ddfc-4063-ab78-2d9ce3190283 req-546e745e-6bf9-4f25-ab72-983dbc874b50 service nova] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Received event network-changed-43989f12-9ac8-484f-a0e9-ad981ecf1fca {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 652.627311] env[62753]: DEBUG nova.compute.manager [req-8b9bacd1-ddfc-4063-ab78-2d9ce3190283 req-546e745e-6bf9-4f25-ab72-983dbc874b50 service nova] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Refreshing instance network info cache due to event network-changed-43989f12-9ac8-484f-a0e9-ad981ecf1fca. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 652.627388] env[62753]: DEBUG oslo_concurrency.lockutils [req-8b9bacd1-ddfc-4063-ab78-2d9ce3190283 req-546e745e-6bf9-4f25-ab72-983dbc874b50 service nova] Acquiring lock "refresh_cache-3c8e531e-033c-4857-9865-b1024e143d44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 652.981808] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Releasing lock "refresh_cache-989db77e-169f-4b3b-a0b9-c4417f4e89c4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 652.982753] env[62753]: DEBUG nova.compute.manager [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 652.982753] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 652.982895] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-97120c3e-c108-442f-8f1b-5462c1312d65 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 652.991565] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4e1080eb-663f-4b15-a7ba-11b5119fdf57 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.018595] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 989db77e-169f-4b3b-a0b9-c4417f4e89c4 could not be found. [ 653.018806] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 653.018984] env[62753]: INFO nova.compute.manager [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 653.019242] env[62753]: DEBUG oslo.service.loopingcall [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 653.020000] env[62753]: DEBUG nova.compute.manager [-] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.020114] env[62753]: DEBUG nova.network.neutron [-] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 653.022124] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.038615] env[62753]: DEBUG nova.network.neutron [-] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.065454] env[62753]: DEBUG nova.network.neutron [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.119720] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.124415] env[62753]: DEBUG nova.network.neutron [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.545273] env[62753]: DEBUG nova.network.neutron [-] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 653.624241] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Releasing lock "refresh_cache-11cf82b0-ea78-40b2-b737-dea8b8703ac3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.624241] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 653.625610] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.626115] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 653.629170] env[62753]: DEBUG oslo_concurrency.lockutils [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Releasing lock "refresh_cache-3c8e531e-033c-4857-9865-b1024e143d44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 653.629655] env[62753]: DEBUG nova.compute.manager [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 653.629935] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 653.630918] env[62753]: DEBUG oslo_concurrency.lockutils [req-8b9bacd1-ddfc-4063-ab78-2d9ce3190283 req-546e745e-6bf9-4f25-ab72-983dbc874b50 service nova] Acquired lock "refresh_cache-3c8e531e-033c-4857-9865-b1024e143d44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 653.631400] env[62753]: DEBUG nova.network.neutron [req-8b9bacd1-ddfc-4063-ab78-2d9ce3190283 req-546e745e-6bf9-4f25-ab72-983dbc874b50 service nova] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Refreshing network info cache for port 43989f12-9ac8-484f-a0e9-ad981ecf1fca {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 653.632444] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cef75f73-fc20-4e1c-89ee-2ea26351739f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.646110] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cea5c0f3-7674-4316-8f76-30c62a98c1f9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.660513] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.674087] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3c8e531e-033c-4857-9865-b1024e143d44 could not be found. [ 653.674438] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 653.674653] env[62753]: INFO nova.compute.manager [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Took 0.04 seconds to destroy the instance on the hypervisor. [ 653.674833] env[62753]: DEBUG oslo.service.loopingcall [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 653.675682] env[62753]: DEBUG nova.compute.manager [-] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 653.675785] env[62753]: DEBUG nova.network.neutron [-] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 653.702303] env[62753]: DEBUG nova.network.neutron [-] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 653.954149] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0808889b-2eca-467c-b18f-751b6c61f34b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.961885] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6502edd-6a67-4644-819c-001d4a1c2f38 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 653.994357] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642e995b-d6f4-4caa-8759-2958539f8151 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.001833] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d1d331d-6ab3-4dd4-9413-acbcba42d90e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 654.015124] env[62753]: DEBUG nova.compute.provider_tree [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 654.049645] env[62753]: INFO nova.compute.manager [-] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Took 1.03 seconds to deallocate network for instance. [ 654.052223] env[62753]: DEBUG nova.compute.claims [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 654.052409] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.156069] env[62753]: DEBUG nova.network.neutron [req-8b9bacd1-ddfc-4063-ab78-2d9ce3190283 req-546e745e-6bf9-4f25-ab72-983dbc874b50 service nova] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 654.165161] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.204632] env[62753]: DEBUG nova.network.neutron [-] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.210434] env[62753]: DEBUG nova.network.neutron [req-8b9bacd1-ddfc-4063-ab78-2d9ce3190283 req-546e745e-6bf9-4f25-ab72-983dbc874b50 service nova] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 654.518743] env[62753]: DEBUG nova.scheduler.client.report [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 654.668498] env[62753]: INFO nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 11cf82b0-ea78-40b2-b737-dea8b8703ac3] Took 1.04 seconds to deallocate network for instance. [ 654.707460] env[62753]: INFO nova.compute.manager [-] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Took 1.03 seconds to deallocate network for instance. [ 654.710825] env[62753]: DEBUG nova.compute.claims [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 654.710825] env[62753]: DEBUG oslo_concurrency.lockutils [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 654.712474] env[62753]: DEBUG oslo_concurrency.lockutils [req-8b9bacd1-ddfc-4063-ab78-2d9ce3190283 req-546e745e-6bf9-4f25-ab72-983dbc874b50 service nova] Releasing lock "refresh_cache-3c8e531e-033c-4857-9865-b1024e143d44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 654.713856] env[62753]: DEBUG nova.compute.manager [req-8b9bacd1-ddfc-4063-ab78-2d9ce3190283 req-546e745e-6bf9-4f25-ab72-983dbc874b50 service nova] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Received event network-vif-deleted-43989f12-9ac8-484f-a0e9-ad981ecf1fca {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 655.024387] env[62753]: DEBUG oslo_concurrency.lockutils [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.533s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 655.024914] env[62753]: DEBUG nova.compute.manager [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 655.027973] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.669s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 655.029657] env[62753]: INFO nova.compute.claims [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 655.539105] env[62753]: DEBUG nova.compute.utils [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 655.542389] env[62753]: DEBUG nova.compute.manager [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 655.542562] env[62753]: DEBUG nova.network.neutron [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 655.604422] env[62753]: DEBUG nova.policy [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'dcc06d0ab77441a3acc61ee3bec59251', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '81a3a38c81794f4dadb25dd0d6c994a8', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 655.719630] env[62753]: INFO nova.scheduler.client.report [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Deleted allocations for instance 11cf82b0-ea78-40b2-b737-dea8b8703ac3 [ 655.993156] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1852ba4-8d59-4bcf-94bc-9c55f590a657 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.001483] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-886f80e7-6996-4f8d-a7f3-a1aa7f315d92 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.036755] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7970b3d-3a25-4d55-9da0-b88e11cbecb1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.043522] env[62753]: DEBUG nova.compute.manager [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 656.051264] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-482c754d-87d8-4898-962a-8ed00f52cde1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 656.066429] env[62753]: DEBUG nova.compute.provider_tree [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 656.071562] env[62753]: DEBUG nova.network.neutron [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Successfully created port: 13eace83-d89c-4c35-aa6d-541fc4278c56 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 656.232023] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "11cf82b0-ea78-40b2-b737-dea8b8703ac3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.472s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 656.569866] env[62753]: DEBUG nova.scheduler.client.report [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 656.734813] env[62753]: DEBUG nova.compute.manager [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 657.070278] env[62753]: DEBUG nova.compute.manager [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 657.074410] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.046s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 657.074410] env[62753]: DEBUG nova.compute.manager [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 657.078713] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.321s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 657.098669] env[62753]: DEBUG oslo_concurrency.lockutils [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "32dae012-7743-4efb-b39b-820c9ed42cc5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.102315] env[62753]: DEBUG nova.virt.hardware [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 657.102550] env[62753]: DEBUG nova.virt.hardware [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 657.102701] env[62753]: DEBUG nova.virt.hardware [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 657.102877] env[62753]: DEBUG nova.virt.hardware [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 657.103039] env[62753]: DEBUG nova.virt.hardware [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 657.105026] env[62753]: DEBUG nova.virt.hardware [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 657.105026] env[62753]: DEBUG nova.virt.hardware [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 657.105026] env[62753]: DEBUG nova.virt.hardware [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 657.105026] env[62753]: DEBUG nova.virt.hardware [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 657.105026] env[62753]: DEBUG nova.virt.hardware [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 657.105201] env[62753]: DEBUG nova.virt.hardware [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 657.105357] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4df2570a-0aed-4d06-bcfb-2e4274e94e69 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.117875] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94e574e7-6d70-4408-a951-d6447d390f66 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 657.261270] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 657.329649] env[62753]: DEBUG nova.compute.manager [req-aa83410a-7122-45bc-a0b3-45344d401543 req-89f1bc61-9ea2-4b68-ba17-745c6b3fb3c8 service nova] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Received event network-changed-13eace83-d89c-4c35-aa6d-541fc4278c56 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 657.331136] env[62753]: DEBUG nova.compute.manager [req-aa83410a-7122-45bc-a0b3-45344d401543 req-89f1bc61-9ea2-4b68-ba17-745c6b3fb3c8 service nova] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Refreshing instance network info cache due to event network-changed-13eace83-d89c-4c35-aa6d-541fc4278c56. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 657.331136] env[62753]: DEBUG oslo_concurrency.lockutils [req-aa83410a-7122-45bc-a0b3-45344d401543 req-89f1bc61-9ea2-4b68-ba17-745c6b3fb3c8 service nova] Acquiring lock "refresh_cache-e730e4f0-fe02-4926-8263-93c8a3065962" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.331136] env[62753]: DEBUG oslo_concurrency.lockutils [req-aa83410a-7122-45bc-a0b3-45344d401543 req-89f1bc61-9ea2-4b68-ba17-745c6b3fb3c8 service nova] Acquired lock "refresh_cache-e730e4f0-fe02-4926-8263-93c8a3065962" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 657.331136] env[62753]: DEBUG nova.network.neutron [req-aa83410a-7122-45bc-a0b3-45344d401543 req-89f1bc61-9ea2-4b68-ba17-745c6b3fb3c8 service nova] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Refreshing network info cache for port 13eace83-d89c-4c35-aa6d-541fc4278c56 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 657.515532] env[62753]: ERROR nova.compute.manager [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 13eace83-d89c-4c35-aa6d-541fc4278c56, please check neutron logs for more information. [ 657.515532] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 657.515532] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.515532] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 657.515532] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.515532] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 657.515532] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.515532] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 657.515532] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.515532] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 657.515532] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.515532] env[62753]: ERROR nova.compute.manager raise self.value [ 657.515532] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.515532] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 657.515532] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.515532] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 657.515979] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.515979] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 657.515979] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 13eace83-d89c-4c35-aa6d-541fc4278c56, please check neutron logs for more information. [ 657.515979] env[62753]: ERROR nova.compute.manager [ 657.515979] env[62753]: Traceback (most recent call last): [ 657.515979] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 657.515979] env[62753]: listener.cb(fileno) [ 657.515979] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.515979] env[62753]: result = function(*args, **kwargs) [ 657.515979] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 657.515979] env[62753]: return func(*args, **kwargs) [ 657.515979] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.515979] env[62753]: raise e [ 657.515979] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.515979] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 657.515979] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.515979] env[62753]: created_port_ids = self._update_ports_for_instance( [ 657.515979] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.515979] env[62753]: with excutils.save_and_reraise_exception(): [ 657.515979] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.515979] env[62753]: self.force_reraise() [ 657.515979] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.515979] env[62753]: raise self.value [ 657.515979] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.515979] env[62753]: updated_port = self._update_port( [ 657.515979] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.515979] env[62753]: _ensure_no_port_binding_failure(port) [ 657.515979] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.515979] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 657.516700] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 13eace83-d89c-4c35-aa6d-541fc4278c56, please check neutron logs for more information. [ 657.516700] env[62753]: Removing descriptor: 16 [ 657.516700] env[62753]: ERROR nova.compute.manager [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 13eace83-d89c-4c35-aa6d-541fc4278c56, please check neutron logs for more information. [ 657.516700] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Traceback (most recent call last): [ 657.516700] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 657.516700] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] yield resources [ 657.516700] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 657.516700] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] self.driver.spawn(context, instance, image_meta, [ 657.516700] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 657.516700] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] self._vmops.spawn(context, instance, image_meta, injected_files, [ 657.516700] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 657.516700] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] vm_ref = self.build_virtual_machine(instance, [ 657.517037] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 657.517037] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] vif_infos = vmwarevif.get_vif_info(self._session, [ 657.517037] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 657.517037] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] for vif in network_info: [ 657.517037] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 657.517037] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] return self._sync_wrapper(fn, *args, **kwargs) [ 657.517037] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 657.517037] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] self.wait() [ 657.517037] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 657.517037] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] self[:] = self._gt.wait() [ 657.517037] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 657.517037] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] return self._exit_event.wait() [ 657.517037] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 657.517374] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] result = hub.switch() [ 657.517374] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 657.517374] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] return self.greenlet.switch() [ 657.517374] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 657.517374] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] result = function(*args, **kwargs) [ 657.517374] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 657.517374] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] return func(*args, **kwargs) [ 657.517374] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 657.517374] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] raise e [ 657.517374] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 657.517374] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] nwinfo = self.network_api.allocate_for_instance( [ 657.517374] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 657.517374] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] created_port_ids = self._update_ports_for_instance( [ 657.517764] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 657.517764] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] with excutils.save_and_reraise_exception(): [ 657.517764] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 657.517764] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] self.force_reraise() [ 657.517764] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 657.517764] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] raise self.value [ 657.517764] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 657.517764] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] updated_port = self._update_port( [ 657.517764] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 657.517764] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] _ensure_no_port_binding_failure(port) [ 657.517764] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 657.517764] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] raise exception.PortBindingFailed(port_id=port['id']) [ 657.518165] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] nova.exception.PortBindingFailed: Binding failed for port 13eace83-d89c-4c35-aa6d-541fc4278c56, please check neutron logs for more information. [ 657.518165] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] [ 657.518165] env[62753]: INFO nova.compute.manager [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Terminating instance [ 657.525104] env[62753]: DEBUG oslo_concurrency.lockutils [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Acquiring lock "refresh_cache-e730e4f0-fe02-4926-8263-93c8a3065962" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 657.584025] env[62753]: DEBUG nova.compute.utils [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 657.584508] env[62753]: DEBUG nova.compute.manager [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 657.584800] env[62753]: DEBUG nova.network.neutron [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 657.631210] env[62753]: DEBUG nova.policy [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '1c51848ff665453c8771692a5b725881', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0470ee02874f4f21b76acffb9a272e94', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 657.702497] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 657.703022] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 657.866623] env[62753]: DEBUG nova.network.neutron [req-aa83410a-7122-45bc-a0b3-45344d401543 req-89f1bc61-9ea2-4b68-ba17-745c6b3fb3c8 service nova] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 657.958211] env[62753]: DEBUG nova.network.neutron [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Successfully created port: ce5b0145-14e4-46d5-90da-42f0588ecc04 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 658.006669] env[62753]: DEBUG nova.network.neutron [req-aa83410a-7122-45bc-a0b3-45344d401543 req-89f1bc61-9ea2-4b68-ba17-745c6b3fb3c8 service nova] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 658.040101] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15071095-23fe-4daf-ae8a-38ee0b4b3982 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.050833] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54a0926e-9eb1-4e05-b852-d09b7a14dee0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.093182] env[62753]: DEBUG nova.compute.manager [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 658.095155] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e10d0509-33e2-48eb-8ee9-d41dcb16f46c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.103090] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecb7502a-22a6-4a42-af20-a1a8d835cc5d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 658.118931] env[62753]: DEBUG nova.compute.provider_tree [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 658.209983] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 658.209983] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Starting heal instance info cache {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 658.209983] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Rebuilding the list of instances to heal {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 658.322340] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Acquiring lock "e260eaf7-98fe-461a-bc69-fc42dfcb79a8" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 658.322675] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Lock "e260eaf7-98fe-461a-bc69-fc42dfcb79a8" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 658.510800] env[62753]: DEBUG oslo_concurrency.lockutils [req-aa83410a-7122-45bc-a0b3-45344d401543 req-89f1bc61-9ea2-4b68-ba17-745c6b3fb3c8 service nova] Releasing lock "refresh_cache-e730e4f0-fe02-4926-8263-93c8a3065962" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 658.511239] env[62753]: DEBUG oslo_concurrency.lockutils [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Acquired lock "refresh_cache-e730e4f0-fe02-4926-8263-93c8a3065962" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 658.511420] env[62753]: DEBUG nova.network.neutron [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 658.623449] env[62753]: DEBUG nova.scheduler.client.report [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 658.713078] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 658.713217] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 658.713340] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 658.713455] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 658.713578] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 658.713766] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Didn't find any instances for network info cache update. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 658.714077] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 658.714077] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 658.714330] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 658.714420] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 658.714475] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 658.714615] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 658.714737] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62753) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 658.714875] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 658.834111] env[62753]: ERROR nova.compute.manager [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ce5b0145-14e4-46d5-90da-42f0588ecc04, please check neutron logs for more information. [ 658.834111] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 658.834111] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.834111] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 658.834111] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.834111] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 658.834111] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.834111] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 658.834111] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.834111] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 658.834111] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.834111] env[62753]: ERROR nova.compute.manager raise self.value [ 658.834111] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.834111] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 658.834111] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.834111] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 658.834572] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.834572] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 658.834572] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ce5b0145-14e4-46d5-90da-42f0588ecc04, please check neutron logs for more information. [ 658.834572] env[62753]: ERROR nova.compute.manager [ 658.834572] env[62753]: Traceback (most recent call last): [ 658.834572] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 658.834572] env[62753]: listener.cb(fileno) [ 658.834572] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 658.834572] env[62753]: result = function(*args, **kwargs) [ 658.834572] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 658.834572] env[62753]: return func(*args, **kwargs) [ 658.834572] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 658.834572] env[62753]: raise e [ 658.834572] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 658.834572] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 658.834572] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 658.834572] env[62753]: created_port_ids = self._update_ports_for_instance( [ 658.834572] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 658.834572] env[62753]: with excutils.save_and_reraise_exception(): [ 658.834572] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 658.834572] env[62753]: self.force_reraise() [ 658.834572] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 658.834572] env[62753]: raise self.value [ 658.834572] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 658.834572] env[62753]: updated_port = self._update_port( [ 658.834572] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 658.834572] env[62753]: _ensure_no_port_binding_failure(port) [ 658.834572] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 658.834572] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 658.835361] env[62753]: nova.exception.PortBindingFailed: Binding failed for port ce5b0145-14e4-46d5-90da-42f0588ecc04, please check neutron logs for more information. [ 658.835361] env[62753]: Removing descriptor: 16 [ 659.034549] env[62753]: DEBUG nova.network.neutron [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.104296] env[62753]: DEBUG nova.compute.manager [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 659.125303] env[62753]: DEBUG nova.network.neutron [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.130281] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.053s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 659.130963] env[62753]: ERROR nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 75747bb4-bc24-4d44-812f-8f0a74badb68, please check neutron logs for more information. [ 659.130963] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Traceback (most recent call last): [ 659.130963] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 659.130963] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] self.driver.spawn(context, instance, image_meta, [ 659.130963] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 659.130963] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] self._vmops.spawn(context, instance, image_meta, injected_files, [ 659.130963] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 659.130963] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] vm_ref = self.build_virtual_machine(instance, [ 659.130963] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 659.130963] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] vif_infos = vmwarevif.get_vif_info(self._session, [ 659.130963] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 659.131396] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] for vif in network_info: [ 659.131396] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 659.131396] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] return self._sync_wrapper(fn, *args, **kwargs) [ 659.131396] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 659.131396] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] self.wait() [ 659.131396] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 659.131396] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] self[:] = self._gt.wait() [ 659.131396] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 659.131396] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] return self._exit_event.wait() [ 659.131396] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 659.131396] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] result = hub.switch() [ 659.131396] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 659.131396] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] return self.greenlet.switch() [ 659.131899] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.131899] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] result = function(*args, **kwargs) [ 659.131899] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 659.131899] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] return func(*args, **kwargs) [ 659.131899] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 659.131899] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] raise e [ 659.131899] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.131899] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] nwinfo = self.network_api.allocate_for_instance( [ 659.131899] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 659.131899] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] created_port_ids = self._update_ports_for_instance( [ 659.131899] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 659.131899] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] with excutils.save_and_reraise_exception(): [ 659.131899] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.132191] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] self.force_reraise() [ 659.132191] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.132191] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] raise self.value [ 659.132191] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 659.132191] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] updated_port = self._update_port( [ 659.132191] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.132191] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] _ensure_no_port_binding_failure(port) [ 659.132191] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.132191] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] raise exception.PortBindingFailed(port_id=port['id']) [ 659.132191] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] nova.exception.PortBindingFailed: Binding failed for port 75747bb4-bc24-4d44-812f-8f0a74badb68, please check neutron logs for more information. [ 659.132191] env[62753]: ERROR nova.compute.manager [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] [ 659.132523] env[62753]: DEBUG nova.compute.utils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Binding failed for port 75747bb4-bc24-4d44-812f-8f0a74badb68, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 659.137087] env[62753]: DEBUG oslo_concurrency.lockutils [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.713s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 659.138601] env[62753]: INFO nova.compute.claims [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 659.141193] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Build of instance 32dae012-7743-4efb-b39b-820c9ed42cc5 was re-scheduled: Binding failed for port 75747bb4-bc24-4d44-812f-8f0a74badb68, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 659.141628] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 659.141877] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "refresh_cache-32dae012-7743-4efb-b39b-820c9ed42cc5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.142036] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquired lock "refresh_cache-32dae012-7743-4efb-b39b-820c9ed42cc5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.142207] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 659.149388] env[62753]: DEBUG nova.virt.hardware [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 659.149520] env[62753]: DEBUG nova.virt.hardware [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 659.149674] env[62753]: DEBUG nova.virt.hardware [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 659.149855] env[62753]: DEBUG nova.virt.hardware [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 659.150008] env[62753]: DEBUG nova.virt.hardware [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 659.150171] env[62753]: DEBUG nova.virt.hardware [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 659.150380] env[62753]: DEBUG nova.virt.hardware [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 659.150537] env[62753]: DEBUG nova.virt.hardware [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 659.150737] env[62753]: DEBUG nova.virt.hardware [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 659.150908] env[62753]: DEBUG nova.virt.hardware [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 659.151095] env[62753]: DEBUG nova.virt.hardware [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 659.152197] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f1e2e34-41fd-4385-9f40-1ed631da391e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.161950] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e23e01-d8eb-4c33-a56a-5c98199feeb1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.177925] env[62753]: ERROR nova.compute.manager [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ce5b0145-14e4-46d5-90da-42f0588ecc04, please check neutron logs for more information. [ 659.177925] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Traceback (most recent call last): [ 659.177925] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 659.177925] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] yield resources [ 659.177925] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 659.177925] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] self.driver.spawn(context, instance, image_meta, [ 659.177925] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 659.177925] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] self._vmops.spawn(context, instance, image_meta, injected_files, [ 659.177925] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 659.177925] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] vm_ref = self.build_virtual_machine(instance, [ 659.177925] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 659.178268] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] vif_infos = vmwarevif.get_vif_info(self._session, [ 659.178268] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 659.178268] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] for vif in network_info: [ 659.178268] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 659.178268] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] return self._sync_wrapper(fn, *args, **kwargs) [ 659.178268] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 659.178268] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] self.wait() [ 659.178268] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 659.178268] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] self[:] = self._gt.wait() [ 659.178268] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 659.178268] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] return self._exit_event.wait() [ 659.178268] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 659.178268] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] current.throw(*self._exc) [ 659.178598] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 659.178598] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] result = function(*args, **kwargs) [ 659.178598] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 659.178598] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] return func(*args, **kwargs) [ 659.178598] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 659.178598] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] raise e [ 659.178598] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 659.178598] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] nwinfo = self.network_api.allocate_for_instance( [ 659.178598] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 659.178598] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] created_port_ids = self._update_ports_for_instance( [ 659.178598] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 659.178598] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] with excutils.save_and_reraise_exception(): [ 659.178598] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 659.178870] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] self.force_reraise() [ 659.178870] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 659.178870] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] raise self.value [ 659.178870] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 659.178870] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] updated_port = self._update_port( [ 659.178870] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 659.178870] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] _ensure_no_port_binding_failure(port) [ 659.178870] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 659.178870] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] raise exception.PortBindingFailed(port_id=port['id']) [ 659.178870] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] nova.exception.PortBindingFailed: Binding failed for port ce5b0145-14e4-46d5-90da-42f0588ecc04, please check neutron logs for more information. [ 659.178870] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] [ 659.178870] env[62753]: INFO nova.compute.manager [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Terminating instance [ 659.180291] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Acquiring lock "refresh_cache-e8f80847-a14d-4abd-af9f-84c62dd1e395" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.180450] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Acquired lock "refresh_cache-e8f80847-a14d-4abd-af9f-84c62dd1e395" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 659.181090] env[62753]: DEBUG nova.network.neutron [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 659.219995] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 659.372206] env[62753]: DEBUG nova.compute.manager [req-10a1be5e-d618-432f-a21d-ef4f93ca6a2c req-ed78c739-0a96-4790-b662-5db86ff7ac25 service nova] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Received event network-vif-deleted-13eace83-d89c-4c35-aa6d-541fc4278c56 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 659.372420] env[62753]: DEBUG nova.compute.manager [req-10a1be5e-d618-432f-a21d-ef4f93ca6a2c req-ed78c739-0a96-4790-b662-5db86ff7ac25 service nova] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Received event network-changed-ce5b0145-14e4-46d5-90da-42f0588ecc04 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 659.372584] env[62753]: DEBUG nova.compute.manager [req-10a1be5e-d618-432f-a21d-ef4f93ca6a2c req-ed78c739-0a96-4790-b662-5db86ff7ac25 service nova] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Refreshing instance network info cache due to event network-changed-ce5b0145-14e4-46d5-90da-42f0588ecc04. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 659.372768] env[62753]: DEBUG oslo_concurrency.lockutils [req-10a1be5e-d618-432f-a21d-ef4f93ca6a2c req-ed78c739-0a96-4790-b662-5db86ff7ac25 service nova] Acquiring lock "refresh_cache-e8f80847-a14d-4abd-af9f-84c62dd1e395" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 659.630786] env[62753]: DEBUG oslo_concurrency.lockutils [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Releasing lock "refresh_cache-e730e4f0-fe02-4926-8263-93c8a3065962" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 659.631232] env[62753]: DEBUG nova.compute.manager [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 659.631425] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 659.632495] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4d1d76f6-5938-4e74-b4ce-53eea94848c9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.640978] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36e698d9-c13d-42e0-a33a-9528fe73980b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 659.666012] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e730e4f0-fe02-4926-8263-93c8a3065962 could not be found. [ 659.666012] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 659.666142] env[62753]: INFO nova.compute.manager [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Took 0.03 seconds to destroy the instance on the hypervisor. [ 659.666751] env[62753]: DEBUG oslo.service.loopingcall [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 659.666751] env[62753]: DEBUG nova.compute.manager [-] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 659.666751] env[62753]: DEBUG nova.network.neutron [-] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 659.669824] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.683295] env[62753]: DEBUG nova.network.neutron [-] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.701638] env[62753]: DEBUG nova.network.neutron [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 659.751312] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 659.792257] env[62753]: DEBUG nova.network.neutron [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.187432] env[62753]: DEBUG nova.network.neutron [-] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.254451] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Releasing lock "refresh_cache-32dae012-7743-4efb-b39b-820c9ed42cc5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.254683] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 660.254863] env[62753]: DEBUG nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 660.255038] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 660.268255] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.294242] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Releasing lock "refresh_cache-e8f80847-a14d-4abd-af9f-84c62dd1e395" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 660.295020] env[62753]: DEBUG nova.compute.manager [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 660.295020] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 660.295408] env[62753]: DEBUG oslo_concurrency.lockutils [req-10a1be5e-d618-432f-a21d-ef4f93ca6a2c req-ed78c739-0a96-4790-b662-5db86ff7ac25 service nova] Acquired lock "refresh_cache-e8f80847-a14d-4abd-af9f-84c62dd1e395" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 660.295655] env[62753]: DEBUG nova.network.neutron [req-10a1be5e-d618-432f-a21d-ef4f93ca6a2c req-ed78c739-0a96-4790-b662-5db86ff7ac25 service nova] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Refreshing network info cache for port ce5b0145-14e4-46d5-90da-42f0588ecc04 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 660.299019] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a739de2f-24cc-4637-ae08-5ffa2108f17f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.305485] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81402487-1482-4c61-9b9a-3c6c51d73d5a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.331211] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e8f80847-a14d-4abd-af9f-84c62dd1e395 could not be found. [ 660.331352] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 660.331551] env[62753]: INFO nova.compute.manager [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Took 0.04 seconds to destroy the instance on the hypervisor. [ 660.331794] env[62753]: DEBUG oslo.service.loopingcall [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 660.334059] env[62753]: DEBUG nova.compute.manager [-] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 660.334163] env[62753]: DEBUG nova.network.neutron [-] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 660.351794] env[62753]: DEBUG nova.network.neutron [-] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.496973] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-561bd7bf-5396-4652-b735-e55827a03436 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.504642] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1fd3bf1-bdd5-4c2d-a74d-3d7cae00dbbf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.535192] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbfac925-ccdd-437f-b83c-52ab1319a7ca {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.542558] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b3ca72d-b0ad-4a07-aabc-f6f011e22085 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 660.555239] env[62753]: DEBUG nova.compute.provider_tree [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 660.689964] env[62753]: INFO nova.compute.manager [-] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Took 1.02 seconds to deallocate network for instance. [ 660.692194] env[62753]: DEBUG nova.compute.claims [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 660.692372] env[62753]: DEBUG oslo_concurrency.lockutils [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 660.769986] env[62753]: DEBUG nova.network.neutron [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.818566] env[62753]: DEBUG nova.network.neutron [req-10a1be5e-d618-432f-a21d-ef4f93ca6a2c req-ed78c739-0a96-4790-b662-5db86ff7ac25 service nova] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 660.854371] env[62753]: DEBUG nova.network.neutron [-] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 660.911710] env[62753]: DEBUG nova.network.neutron [req-10a1be5e-d618-432f-a21d-ef4f93ca6a2c req-ed78c739-0a96-4790-b662-5db86ff7ac25 service nova] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 661.058470] env[62753]: DEBUG nova.scheduler.client.report [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 661.275531] env[62753]: INFO nova.compute.manager [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Took 1.02 seconds to deallocate network for instance. [ 661.356875] env[62753]: INFO nova.compute.manager [-] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Took 1.02 seconds to deallocate network for instance. [ 661.359246] env[62753]: DEBUG nova.compute.claims [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 661.359388] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 661.414806] env[62753]: DEBUG oslo_concurrency.lockutils [req-10a1be5e-d618-432f-a21d-ef4f93ca6a2c req-ed78c739-0a96-4790-b662-5db86ff7ac25 service nova] Releasing lock "refresh_cache-e8f80847-a14d-4abd-af9f-84c62dd1e395" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 661.415121] env[62753]: DEBUG nova.compute.manager [req-10a1be5e-d618-432f-a21d-ef4f93ca6a2c req-ed78c739-0a96-4790-b662-5db86ff7ac25 service nova] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Received event network-vif-deleted-ce5b0145-14e4-46d5-90da-42f0588ecc04 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 661.563395] env[62753]: DEBUG oslo_concurrency.lockutils [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.426s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 661.563912] env[62753]: DEBUG nova.compute.manager [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 661.566412] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.075s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 661.569230] env[62753]: INFO nova.compute.claims [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 662.074385] env[62753]: DEBUG nova.compute.utils [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 662.080019] env[62753]: DEBUG nova.compute.manager [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 662.080019] env[62753]: DEBUG nova.network.neutron [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 662.144620] env[62753]: DEBUG nova.policy [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd86dfa341c7a41938421b65e5d8d35b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fdd163b294804fffbf0e92873611d5ea', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 662.305801] env[62753]: INFO nova.scheduler.client.report [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Deleted allocations for instance 32dae012-7743-4efb-b39b-820c9ed42cc5 [ 662.582493] env[62753]: DEBUG nova.compute.manager [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 662.607742] env[62753]: DEBUG nova.network.neutron [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Successfully created port: ad3348fb-2ed5-4d71-adfb-edfbeb299328 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 662.815183] env[62753]: DEBUG oslo_concurrency.lockutils [None req-09c8e63f-ae1b-404e-8623-c20f996c8ff5 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "32dae012-7743-4efb-b39b-820c9ed42cc5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.988s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.816620] env[62753]: DEBUG oslo_concurrency.lockutils [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "32dae012-7743-4efb-b39b-820c9ed42cc5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 5.718s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.816980] env[62753]: DEBUG oslo_concurrency.lockutils [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "32dae012-7743-4efb-b39b-820c9ed42cc5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 662.817257] env[62753]: DEBUG oslo_concurrency.lockutils [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "32dae012-7743-4efb-b39b-820c9ed42cc5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 662.817433] env[62753]: DEBUG oslo_concurrency.lockutils [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "32dae012-7743-4efb-b39b-820c9ed42cc5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 662.819214] env[62753]: INFO nova.compute.manager [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Terminating instance [ 662.820986] env[62753]: DEBUG oslo_concurrency.lockutils [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "refresh_cache-32dae012-7743-4efb-b39b-820c9ed42cc5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 662.821167] env[62753]: DEBUG oslo_concurrency.lockutils [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquired lock "refresh_cache-32dae012-7743-4efb-b39b-820c9ed42cc5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 662.821336] env[62753]: DEBUG nova.network.neutron [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 662.971027] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1645af72-0c68-4e97-acea-8acf51143333 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 662.979216] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea3b8e20-91b1-4042-940c-e64403510b21 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.014590] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a5774f-27bc-48eb-9eb1-5c1ebaa11b8f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.022783] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-919152a2-bac0-49d6-9b85-46711b2940d8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.037535] env[62753]: DEBUG nova.compute.provider_tree [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 663.320278] env[62753]: DEBUG nova.compute.manager [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 663.344272] env[62753]: DEBUG nova.network.neutron [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 663.419980] env[62753]: DEBUG nova.network.neutron [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 663.541358] env[62753]: DEBUG nova.scheduler.client.report [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 663.592750] env[62753]: DEBUG nova.compute.manager [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 663.628993] env[62753]: DEBUG nova.virt.hardware [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 663.629252] env[62753]: DEBUG nova.virt.hardware [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 663.629410] env[62753]: DEBUG nova.virt.hardware [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 663.629787] env[62753]: DEBUG nova.virt.hardware [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 663.629787] env[62753]: DEBUG nova.virt.hardware [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 663.629947] env[62753]: DEBUG nova.virt.hardware [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 663.630141] env[62753]: DEBUG nova.virt.hardware [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 663.630258] env[62753]: DEBUG nova.virt.hardware [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 663.630389] env[62753]: DEBUG nova.virt.hardware [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 663.630590] env[62753]: DEBUG nova.virt.hardware [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 663.630747] env[62753]: DEBUG nova.virt.hardware [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 663.631648] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bc3339b-9716-4c3e-9cbd-7c08698af6e8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.642682] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1cbe2b7-1618-4538-bb04-fcfdfcb87206 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.796413] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "7831fd59-7a6d-4f6c-9f46-8cb1e2661c17" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.796653] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "7831fd59-7a6d-4f6c-9f46-8cb1e2661c17" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.830688] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.830936] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 663.833916] env[62753]: DEBUG nova.compute.manager [req-0d55230e-5d24-4aee-b038-402aaf6527c8 req-25f72253-56e3-470b-801e-14f35517afbd service nova] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Received event network-changed-ad3348fb-2ed5-4d71-adfb-edfbeb299328 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 663.834117] env[62753]: DEBUG nova.compute.manager [req-0d55230e-5d24-4aee-b038-402aaf6527c8 req-25f72253-56e3-470b-801e-14f35517afbd service nova] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Refreshing instance network info cache due to event network-changed-ad3348fb-2ed5-4d71-adfb-edfbeb299328. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 663.834319] env[62753]: DEBUG oslo_concurrency.lockutils [req-0d55230e-5d24-4aee-b038-402aaf6527c8 req-25f72253-56e3-470b-801e-14f35517afbd service nova] Acquiring lock "refresh_cache-3d13009d-5e65-43e1-87b3-3cb3cfd014e6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 663.834456] env[62753]: DEBUG oslo_concurrency.lockutils [req-0d55230e-5d24-4aee-b038-402aaf6527c8 req-25f72253-56e3-470b-801e-14f35517afbd service nova] Acquired lock "refresh_cache-3d13009d-5e65-43e1-87b3-3cb3cfd014e6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 663.834733] env[62753]: DEBUG nova.network.neutron [req-0d55230e-5d24-4aee-b038-402aaf6527c8 req-25f72253-56e3-470b-801e-14f35517afbd service nova] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Refreshing network info cache for port ad3348fb-2ed5-4d71-adfb-edfbeb299328 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 663.853517] env[62753]: DEBUG oslo_concurrency.lockutils [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 663.922828] env[62753]: DEBUG oslo_concurrency.lockutils [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Releasing lock "refresh_cache-32dae012-7743-4efb-b39b-820c9ed42cc5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 663.923322] env[62753]: DEBUG nova.compute.manager [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 663.923549] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 663.923802] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5d1889e3-3715-4b7c-ae07-f2f39e032110 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.932862] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6665965-6cc3-4214-8112-e28b1b4bb62b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 663.956907] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 32dae012-7743-4efb-b39b-820c9ed42cc5 could not be found. [ 663.956907] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 663.956907] env[62753]: INFO nova.compute.manager [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Took 0.03 seconds to destroy the instance on the hypervisor. [ 663.956907] env[62753]: DEBUG oslo.service.loopingcall [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 663.956907] env[62753]: DEBUG nova.compute.manager [-] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 663.956907] env[62753]: DEBUG nova.network.neutron [-] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 663.976922] env[62753]: DEBUG nova.network.neutron [-] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.018484] env[62753]: ERROR nova.compute.manager [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ad3348fb-2ed5-4d71-adfb-edfbeb299328, please check neutron logs for more information. [ 664.018484] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 664.018484] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.018484] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 664.018484] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.018484] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 664.018484] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.018484] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 664.018484] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.018484] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 664.018484] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.018484] env[62753]: ERROR nova.compute.manager raise self.value [ 664.018484] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.018484] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 664.018484] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.018484] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 664.018882] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.018882] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 664.018882] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ad3348fb-2ed5-4d71-adfb-edfbeb299328, please check neutron logs for more information. [ 664.018882] env[62753]: ERROR nova.compute.manager [ 664.018882] env[62753]: Traceback (most recent call last): [ 664.018882] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 664.018882] env[62753]: listener.cb(fileno) [ 664.018882] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.018882] env[62753]: result = function(*args, **kwargs) [ 664.018882] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 664.018882] env[62753]: return func(*args, **kwargs) [ 664.018882] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.018882] env[62753]: raise e [ 664.018882] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.018882] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 664.018882] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.018882] env[62753]: created_port_ids = self._update_ports_for_instance( [ 664.018882] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.018882] env[62753]: with excutils.save_and_reraise_exception(): [ 664.018882] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.018882] env[62753]: self.force_reraise() [ 664.018882] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.018882] env[62753]: raise self.value [ 664.018882] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.018882] env[62753]: updated_port = self._update_port( [ 664.018882] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.018882] env[62753]: _ensure_no_port_binding_failure(port) [ 664.018882] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.018882] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 664.019533] env[62753]: nova.exception.PortBindingFailed: Binding failed for port ad3348fb-2ed5-4d71-adfb-edfbeb299328, please check neutron logs for more information. [ 664.019533] env[62753]: Removing descriptor: 14 [ 664.019804] env[62753]: ERROR nova.compute.manager [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ad3348fb-2ed5-4d71-adfb-edfbeb299328, please check neutron logs for more information. [ 664.019804] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Traceback (most recent call last): [ 664.019804] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 664.019804] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] yield resources [ 664.019804] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 664.019804] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] self.driver.spawn(context, instance, image_meta, [ 664.019804] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 664.019804] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 664.019804] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 664.019804] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] vm_ref = self.build_virtual_machine(instance, [ 664.019804] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 664.020244] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] vif_infos = vmwarevif.get_vif_info(self._session, [ 664.020244] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 664.020244] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] for vif in network_info: [ 664.020244] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 664.020244] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] return self._sync_wrapper(fn, *args, **kwargs) [ 664.020244] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 664.020244] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] self.wait() [ 664.020244] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 664.020244] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] self[:] = self._gt.wait() [ 664.020244] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 664.020244] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] return self._exit_event.wait() [ 664.020244] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 664.020244] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] result = hub.switch() [ 664.020774] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 664.020774] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] return self.greenlet.switch() [ 664.020774] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 664.020774] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] result = function(*args, **kwargs) [ 664.020774] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 664.020774] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] return func(*args, **kwargs) [ 664.020774] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 664.020774] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] raise e [ 664.020774] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 664.020774] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] nwinfo = self.network_api.allocate_for_instance( [ 664.020774] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 664.020774] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] created_port_ids = self._update_ports_for_instance( [ 664.020774] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 664.021266] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] with excutils.save_and_reraise_exception(): [ 664.021266] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 664.021266] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] self.force_reraise() [ 664.021266] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 664.021266] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] raise self.value [ 664.021266] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 664.021266] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] updated_port = self._update_port( [ 664.021266] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 664.021266] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] _ensure_no_port_binding_failure(port) [ 664.021266] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 664.021266] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] raise exception.PortBindingFailed(port_id=port['id']) [ 664.021266] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] nova.exception.PortBindingFailed: Binding failed for port ad3348fb-2ed5-4d71-adfb-edfbeb299328, please check neutron logs for more information. [ 664.021266] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] [ 664.021785] env[62753]: INFO nova.compute.manager [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Terminating instance [ 664.021785] env[62753]: DEBUG oslo_concurrency.lockutils [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Acquiring lock "refresh_cache-3d13009d-5e65-43e1-87b3-3cb3cfd014e6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 664.047569] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.481s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 664.048052] env[62753]: DEBUG nova.compute.manager [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 664.050539] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.599s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 664.052395] env[62753]: INFO nova.compute.claims [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 664.351034] env[62753]: DEBUG nova.network.neutron [req-0d55230e-5d24-4aee-b038-402aaf6527c8 req-25f72253-56e3-470b-801e-14f35517afbd service nova] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 664.397205] env[62753]: DEBUG nova.network.neutron [req-0d55230e-5d24-4aee-b038-402aaf6527c8 req-25f72253-56e3-470b-801e-14f35517afbd service nova] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.479587] env[62753]: DEBUG nova.network.neutron [-] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 664.556769] env[62753]: DEBUG nova.compute.utils [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 664.561084] env[62753]: DEBUG nova.compute.manager [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 664.561084] env[62753]: DEBUG nova.network.neutron [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 664.642460] env[62753]: DEBUG nova.policy [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '43e8ee192375459993b47402a77d6c33', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2931f3a62680431cb570810817bacdb9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 664.900104] env[62753]: DEBUG oslo_concurrency.lockutils [req-0d55230e-5d24-4aee-b038-402aaf6527c8 req-25f72253-56e3-470b-801e-14f35517afbd service nova] Releasing lock "refresh_cache-3d13009d-5e65-43e1-87b3-3cb3cfd014e6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 664.900442] env[62753]: DEBUG oslo_concurrency.lockutils [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Acquired lock "refresh_cache-3d13009d-5e65-43e1-87b3-3cb3cfd014e6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 664.900701] env[62753]: DEBUG nova.network.neutron [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 664.982500] env[62753]: INFO nova.compute.manager [-] [instance: 32dae012-7743-4efb-b39b-820c9ed42cc5] Took 1.03 seconds to deallocate network for instance. [ 665.063017] env[62753]: DEBUG nova.compute.manager [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 665.079235] env[62753]: DEBUG nova.network.neutron [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Successfully created port: c2fd8ca2-0060-40ba-a71d-4804b9f9e9f8 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 665.418013] env[62753]: DEBUG nova.network.neutron [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 665.426836] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f1ce5e9-4e7a-41c9-bc71-fe11456e7acf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.434148] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fc575e6-48d6-4dd3-b12d-9ad41f8dbc71 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.465974] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11ba5a8b-10fa-4226-be20-83e1732ab183 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.473050] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c450666d-a702-436a-bcc6-ba19c2c52c54 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.477274] env[62753]: DEBUG nova.network.neutron [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 665.490103] env[62753]: DEBUG nova.compute.provider_tree [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 665.888019] env[62753]: DEBUG nova.compute.manager [req-6229d472-22be-463f-a87c-2a422926bbb2 req-27a3b9bb-2034-49c4-a6ba-6c4fa603ca67 service nova] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Received event network-vif-deleted-ad3348fb-2ed5-4d71-adfb-edfbeb299328 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 665.980783] env[62753]: DEBUG oslo_concurrency.lockutils [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Releasing lock "refresh_cache-3d13009d-5e65-43e1-87b3-3cb3cfd014e6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 665.981233] env[62753]: DEBUG nova.compute.manager [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 665.981426] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 665.981769] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f5d0ffd7-4046-43f7-aff4-b816ad8db240 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 665.992609] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4120a514-1e57-4853-af4b-a727df50bd42 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.005923] env[62753]: DEBUG nova.scheduler.client.report [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 666.013380] env[62753]: DEBUG oslo_concurrency.lockutils [None req-51ce07f6-4369-423c-bd81-16379ff87fa1 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "32dae012-7743-4efb-b39b-820c9ed42cc5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.197s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.024847] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3d13009d-5e65-43e1-87b3-3cb3cfd014e6 could not be found. [ 666.026657] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 666.026657] env[62753]: INFO nova.compute.manager [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Took 0.04 seconds to destroy the instance on the hypervisor. [ 666.026657] env[62753]: DEBUG oslo.service.loopingcall [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 666.026657] env[62753]: DEBUG nova.compute.manager [-] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 666.026657] env[62753]: DEBUG nova.network.neutron [-] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 666.040622] env[62753]: DEBUG nova.network.neutron [-] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 666.069454] env[62753]: DEBUG nova.compute.manager [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 666.097494] env[62753]: DEBUG nova.virt.hardware [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 666.097635] env[62753]: DEBUG nova.virt.hardware [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 666.097747] env[62753]: DEBUG nova.virt.hardware [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 666.097921] env[62753]: DEBUG nova.virt.hardware [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 666.098077] env[62753]: DEBUG nova.virt.hardware [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 666.098246] env[62753]: DEBUG nova.virt.hardware [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 666.098455] env[62753]: DEBUG nova.virt.hardware [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 666.098746] env[62753]: DEBUG nova.virt.hardware [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 666.098933] env[62753]: DEBUG nova.virt.hardware [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 666.099111] env[62753]: DEBUG nova.virt.hardware [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 666.099285] env[62753]: DEBUG nova.virt.hardware [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 666.100167] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b9d718b-9214-4fdc-bccc-6ad73ba72425 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.108752] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f1bb4cd-2952-4b6a-9db1-df86cf55b497 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 666.423718] env[62753]: ERROR nova.compute.manager [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c2fd8ca2-0060-40ba-a71d-4804b9f9e9f8, please check neutron logs for more information. [ 666.423718] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 666.423718] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.423718] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 666.423718] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 666.423718] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 666.423718] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 666.423718] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 666.423718] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.423718] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 666.423718] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.423718] env[62753]: ERROR nova.compute.manager raise self.value [ 666.423718] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 666.423718] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 666.423718] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.423718] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 666.424342] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.424342] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 666.424342] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c2fd8ca2-0060-40ba-a71d-4804b9f9e9f8, please check neutron logs for more information. [ 666.424342] env[62753]: ERROR nova.compute.manager [ 666.424342] env[62753]: Traceback (most recent call last): [ 666.424342] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 666.424342] env[62753]: listener.cb(fileno) [ 666.424342] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.424342] env[62753]: result = function(*args, **kwargs) [ 666.424342] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 666.424342] env[62753]: return func(*args, **kwargs) [ 666.424342] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.424342] env[62753]: raise e [ 666.424342] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.424342] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 666.424342] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 666.424342] env[62753]: created_port_ids = self._update_ports_for_instance( [ 666.424342] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 666.424342] env[62753]: with excutils.save_and_reraise_exception(): [ 666.424342] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.424342] env[62753]: self.force_reraise() [ 666.424342] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.424342] env[62753]: raise self.value [ 666.424342] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 666.424342] env[62753]: updated_port = self._update_port( [ 666.424342] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.424342] env[62753]: _ensure_no_port_binding_failure(port) [ 666.424342] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.424342] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 666.425252] env[62753]: nova.exception.PortBindingFailed: Binding failed for port c2fd8ca2-0060-40ba-a71d-4804b9f9e9f8, please check neutron logs for more information. [ 666.425252] env[62753]: Removing descriptor: 14 [ 666.425511] env[62753]: ERROR nova.compute.manager [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c2fd8ca2-0060-40ba-a71d-4804b9f9e9f8, please check neutron logs for more information. [ 666.425511] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Traceback (most recent call last): [ 666.425511] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 666.425511] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] yield resources [ 666.425511] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 666.425511] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] self.driver.spawn(context, instance, image_meta, [ 666.425511] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 666.425511] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 666.425511] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 666.425511] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] vm_ref = self.build_virtual_machine(instance, [ 666.425511] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 666.425801] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] vif_infos = vmwarevif.get_vif_info(self._session, [ 666.425801] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 666.425801] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] for vif in network_info: [ 666.425801] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 666.425801] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] return self._sync_wrapper(fn, *args, **kwargs) [ 666.425801] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 666.425801] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] self.wait() [ 666.425801] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 666.425801] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] self[:] = self._gt.wait() [ 666.425801] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 666.425801] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] return self._exit_event.wait() [ 666.425801] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 666.425801] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] result = hub.switch() [ 666.426187] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 666.426187] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] return self.greenlet.switch() [ 666.426187] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 666.426187] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] result = function(*args, **kwargs) [ 666.426187] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 666.426187] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] return func(*args, **kwargs) [ 666.426187] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 666.426187] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] raise e [ 666.426187] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 666.426187] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] nwinfo = self.network_api.allocate_for_instance( [ 666.426187] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 666.426187] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] created_port_ids = self._update_ports_for_instance( [ 666.426187] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 666.426527] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] with excutils.save_and_reraise_exception(): [ 666.426527] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 666.426527] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] self.force_reraise() [ 666.426527] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 666.426527] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] raise self.value [ 666.426527] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 666.426527] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] updated_port = self._update_port( [ 666.426527] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 666.426527] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] _ensure_no_port_binding_failure(port) [ 666.426527] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 666.426527] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] raise exception.PortBindingFailed(port_id=port['id']) [ 666.426527] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] nova.exception.PortBindingFailed: Binding failed for port c2fd8ca2-0060-40ba-a71d-4804b9f9e9f8, please check neutron logs for more information. [ 666.426527] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] [ 666.426938] env[62753]: INFO nova.compute.manager [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Terminating instance [ 666.428727] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Acquiring lock "refresh_cache-b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 666.428955] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Acquired lock "refresh_cache-b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 666.429169] env[62753]: DEBUG nova.network.neutron [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 666.513687] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.463s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 666.514219] env[62753]: DEBUG nova.compute.manager [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 666.517183] env[62753]: DEBUG oslo_concurrency.lockutils [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.318s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 666.543391] env[62753]: DEBUG nova.network.neutron [-] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 666.946917] env[62753]: DEBUG nova.network.neutron [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.023978] env[62753]: DEBUG nova.compute.utils [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 667.026077] env[62753]: DEBUG nova.network.neutron [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 667.026890] env[62753]: DEBUG nova.compute.manager [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 667.027481] env[62753]: DEBUG nova.network.neutron [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 667.046366] env[62753]: INFO nova.compute.manager [-] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Took 1.02 seconds to deallocate network for instance. [ 667.049037] env[62753]: DEBUG nova.compute.claims [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 667.049209] env[62753]: DEBUG oslo_concurrency.lockutils [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 667.085376] env[62753]: DEBUG nova.policy [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4dbc63095b0464fa0d2de86e3cf170e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5636da907ab343f9b42dbbd903d32283', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 667.411026] env[62753]: DEBUG nova.network.neutron [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Successfully created port: 8bf56dd9-d810-4edc-ab37-c57050778d58 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 667.441639] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c0b52e55-7deb-43a5-a152-599f4f052c2e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.451155] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b67882ce-3f12-48d8-b5d8-769d35bf8c82 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.482392] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-08ff33cc-1f42-42d6-bdfc-ad4683275545 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.489718] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e88c7940-74f3-4661-9b08-95ed568c830d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.504739] env[62753]: DEBUG nova.compute.provider_tree [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 667.529237] env[62753]: DEBUG nova.compute.manager [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 667.535208] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Releasing lock "refresh_cache-b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 667.535605] env[62753]: DEBUG nova.compute.manager [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 667.535984] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 667.536449] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-b0dcf6a1-4be4-4bb4-9c7f-39ce8f771b84 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.549184] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f002be5-7144-405e-88e9-4c0ea2a88f52 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 667.577241] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a could not be found. [ 667.577504] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 667.577680] env[62753]: INFO nova.compute.manager [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 667.578106] env[62753]: DEBUG oslo.service.loopingcall [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 667.578357] env[62753]: DEBUG nova.compute.manager [-] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 667.578447] env[62753]: DEBUG nova.network.neutron [-] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 667.597406] env[62753]: DEBUG nova.network.neutron [-] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 667.951815] env[62753]: DEBUG nova.compute.manager [req-42d1f02e-e656-452e-a933-88c7fdcbd715 req-47b979b7-352a-4ec8-9e0e-5ee0e94626b2 service nova] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Received event network-changed-c2fd8ca2-0060-40ba-a71d-4804b9f9e9f8 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 667.952125] env[62753]: DEBUG nova.compute.manager [req-42d1f02e-e656-452e-a933-88c7fdcbd715 req-47b979b7-352a-4ec8-9e0e-5ee0e94626b2 service nova] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Refreshing instance network info cache due to event network-changed-c2fd8ca2-0060-40ba-a71d-4804b9f9e9f8. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 667.952354] env[62753]: DEBUG oslo_concurrency.lockutils [req-42d1f02e-e656-452e-a933-88c7fdcbd715 req-47b979b7-352a-4ec8-9e0e-5ee0e94626b2 service nova] Acquiring lock "refresh_cache-b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 667.952495] env[62753]: DEBUG oslo_concurrency.lockutils [req-42d1f02e-e656-452e-a933-88c7fdcbd715 req-47b979b7-352a-4ec8-9e0e-5ee0e94626b2 service nova] Acquired lock "refresh_cache-b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 667.953017] env[62753]: DEBUG nova.network.neutron [req-42d1f02e-e656-452e-a933-88c7fdcbd715 req-47b979b7-352a-4ec8-9e0e-5ee0e94626b2 service nova] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Refreshing network info cache for port c2fd8ca2-0060-40ba-a71d-4804b9f9e9f8 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 668.007540] env[62753]: DEBUG nova.scheduler.client.report [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 668.100378] env[62753]: DEBUG nova.network.neutron [-] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.474894] env[62753]: DEBUG nova.network.neutron [req-42d1f02e-e656-452e-a933-88c7fdcbd715 req-47b979b7-352a-4ec8-9e0e-5ee0e94626b2 service nova] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 668.516341] env[62753]: DEBUG oslo_concurrency.lockutils [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.999s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 668.516994] env[62753]: ERROR nova.compute.manager [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e72e4a03-e91c-4040-92ce-0c6ed825b671, please check neutron logs for more information. [ 668.516994] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Traceback (most recent call last): [ 668.516994] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 668.516994] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] self.driver.spawn(context, instance, image_meta, [ 668.516994] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 668.516994] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] self._vmops.spawn(context, instance, image_meta, injected_files, [ 668.516994] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 668.516994] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] vm_ref = self.build_virtual_machine(instance, [ 668.516994] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 668.516994] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] vif_infos = vmwarevif.get_vif_info(self._session, [ 668.516994] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 668.517296] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] for vif in network_info: [ 668.517296] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 668.517296] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] return self._sync_wrapper(fn, *args, **kwargs) [ 668.517296] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 668.517296] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] self.wait() [ 668.517296] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 668.517296] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] self[:] = self._gt.wait() [ 668.517296] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 668.517296] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] return self._exit_event.wait() [ 668.517296] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 668.517296] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] result = hub.switch() [ 668.517296] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 668.517296] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] return self.greenlet.switch() [ 668.517583] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.517583] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] result = function(*args, **kwargs) [ 668.517583] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 668.517583] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] return func(*args, **kwargs) [ 668.517583] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.517583] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] raise e [ 668.517583] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.517583] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] nwinfo = self.network_api.allocate_for_instance( [ 668.517583] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.517583] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] created_port_ids = self._update_ports_for_instance( [ 668.517583] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.517583] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] with excutils.save_and_reraise_exception(): [ 668.517583] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.517955] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] self.force_reraise() [ 668.517955] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.517955] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] raise self.value [ 668.517955] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.517955] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] updated_port = self._update_port( [ 668.517955] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.517955] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] _ensure_no_port_binding_failure(port) [ 668.517955] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.517955] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] raise exception.PortBindingFailed(port_id=port['id']) [ 668.517955] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] nova.exception.PortBindingFailed: Binding failed for port e72e4a03-e91c-4040-92ce-0c6ed825b671, please check neutron logs for more information. [ 668.517955] env[62753]: ERROR nova.compute.manager [instance: 2d028151-3d84-4195-9f93-28287dbfda09] [ 668.518261] env[62753]: DEBUG nova.compute.utils [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Binding failed for port e72e4a03-e91c-4040-92ce-0c6ed825b671, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 668.518908] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.466s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 668.522170] env[62753]: DEBUG nova.compute.manager [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Build of instance 2d028151-3d84-4195-9f93-28287dbfda09 was re-scheduled: Binding failed for port e72e4a03-e91c-4040-92ce-0c6ed825b671, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 668.522633] env[62753]: DEBUG nova.compute.manager [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 668.522858] env[62753]: DEBUG oslo_concurrency.lockutils [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Acquiring lock "refresh_cache-2d028151-3d84-4195-9f93-28287dbfda09" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.523030] env[62753]: DEBUG oslo_concurrency.lockutils [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Acquired lock "refresh_cache-2d028151-3d84-4195-9f93-28287dbfda09" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.523217] env[62753]: DEBUG nova.network.neutron [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 668.524677] env[62753]: DEBUG nova.network.neutron [req-42d1f02e-e656-452e-a933-88c7fdcbd715 req-47b979b7-352a-4ec8-9e0e-5ee0e94626b2 service nova] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 668.542360] env[62753]: DEBUG nova.compute.manager [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 668.569513] env[62753]: DEBUG nova.virt.hardware [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 668.569749] env[62753]: DEBUG nova.virt.hardware [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 668.569904] env[62753]: DEBUG nova.virt.hardware [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 668.570102] env[62753]: DEBUG nova.virt.hardware [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 668.570275] env[62753]: DEBUG nova.virt.hardware [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 668.570395] env[62753]: DEBUG nova.virt.hardware [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 668.570604] env[62753]: DEBUG nova.virt.hardware [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 668.570843] env[62753]: DEBUG nova.virt.hardware [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 668.571052] env[62753]: DEBUG nova.virt.hardware [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 668.571221] env[62753]: DEBUG nova.virt.hardware [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 668.571391] env[62753]: DEBUG nova.virt.hardware [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 668.572487] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f78ed6a8-1725-446c-bef2-724d24ef99cf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.576155] env[62753]: ERROR nova.compute.manager [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8bf56dd9-d810-4edc-ab37-c57050778d58, please check neutron logs for more information. [ 668.576155] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 668.576155] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.576155] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 668.576155] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.576155] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 668.576155] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.576155] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 668.576155] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.576155] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 668.576155] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.576155] env[62753]: ERROR nova.compute.manager raise self.value [ 668.576155] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.576155] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 668.576155] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.576155] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 668.576618] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.576618] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 668.576618] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8bf56dd9-d810-4edc-ab37-c57050778d58, please check neutron logs for more information. [ 668.576618] env[62753]: ERROR nova.compute.manager [ 668.576618] env[62753]: Traceback (most recent call last): [ 668.576618] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 668.576618] env[62753]: listener.cb(fileno) [ 668.576618] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.576618] env[62753]: result = function(*args, **kwargs) [ 668.576618] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 668.576618] env[62753]: return func(*args, **kwargs) [ 668.576618] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.576618] env[62753]: raise e [ 668.576618] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.576618] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 668.576618] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.576618] env[62753]: created_port_ids = self._update_ports_for_instance( [ 668.576618] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.576618] env[62753]: with excutils.save_and_reraise_exception(): [ 668.576618] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.576618] env[62753]: self.force_reraise() [ 668.576618] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.576618] env[62753]: raise self.value [ 668.576618] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.576618] env[62753]: updated_port = self._update_port( [ 668.576618] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.576618] env[62753]: _ensure_no_port_binding_failure(port) [ 668.576618] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.576618] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 668.578323] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 8bf56dd9-d810-4edc-ab37-c57050778d58, please check neutron logs for more information. [ 668.578323] env[62753]: Removing descriptor: 14 [ 668.580414] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8b4f111-c76a-48c9-afd5-1bed0d204bcc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 668.596053] env[62753]: ERROR nova.compute.manager [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8bf56dd9-d810-4edc-ab37-c57050778d58, please check neutron logs for more information. [ 668.596053] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Traceback (most recent call last): [ 668.596053] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 668.596053] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] yield resources [ 668.596053] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 668.596053] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] self.driver.spawn(context, instance, image_meta, [ 668.596053] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 668.596053] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 668.596053] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 668.596053] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] vm_ref = self.build_virtual_machine(instance, [ 668.596053] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 668.596375] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] vif_infos = vmwarevif.get_vif_info(self._session, [ 668.596375] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 668.596375] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] for vif in network_info: [ 668.596375] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 668.596375] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] return self._sync_wrapper(fn, *args, **kwargs) [ 668.596375] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 668.596375] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] self.wait() [ 668.596375] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 668.596375] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] self[:] = self._gt.wait() [ 668.596375] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 668.596375] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] return self._exit_event.wait() [ 668.596375] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 668.596375] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] current.throw(*self._exc) [ 668.596659] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 668.596659] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] result = function(*args, **kwargs) [ 668.596659] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 668.596659] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] return func(*args, **kwargs) [ 668.596659] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 668.596659] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] raise e [ 668.596659] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 668.596659] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] nwinfo = self.network_api.allocate_for_instance( [ 668.596659] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 668.596659] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] created_port_ids = self._update_ports_for_instance( [ 668.596659] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 668.596659] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] with excutils.save_and_reraise_exception(): [ 668.596659] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 668.596945] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] self.force_reraise() [ 668.596945] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 668.596945] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] raise self.value [ 668.596945] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 668.596945] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] updated_port = self._update_port( [ 668.596945] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 668.596945] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] _ensure_no_port_binding_failure(port) [ 668.596945] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 668.596945] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] raise exception.PortBindingFailed(port_id=port['id']) [ 668.596945] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] nova.exception.PortBindingFailed: Binding failed for port 8bf56dd9-d810-4edc-ab37-c57050778d58, please check neutron logs for more information. [ 668.596945] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] [ 668.596945] env[62753]: INFO nova.compute.manager [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Terminating instance [ 668.598663] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "refresh_cache-a408799d-29a9-442f-9769-8f8438af8ae0" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 668.598820] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "refresh_cache-a408799d-29a9-442f-9769-8f8438af8ae0" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 668.598982] env[62753]: DEBUG nova.network.neutron [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 668.602718] env[62753]: INFO nova.compute.manager [-] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Took 1.02 seconds to deallocate network for instance. [ 668.604438] env[62753]: DEBUG nova.compute.claims [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 668.604609] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 669.588633] env[62753]: DEBUG oslo_concurrency.lockutils [req-42d1f02e-e656-452e-a933-88c7fdcbd715 req-47b979b7-352a-4ec8-9e0e-5ee0e94626b2 service nova] Releasing lock "refresh_cache-b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 669.588929] env[62753]: DEBUG nova.compute.manager [req-42d1f02e-e656-452e-a933-88c7fdcbd715 req-47b979b7-352a-4ec8-9e0e-5ee0e94626b2 service nova] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Received event network-vif-deleted-c2fd8ca2-0060-40ba-a71d-4804b9f9e9f8 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 669.605149] env[62753]: DEBUG nova.network.neutron [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.609373] env[62753]: DEBUG nova.network.neutron [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 669.675868] env[62753]: DEBUG nova.network.neutron [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.698687] env[62753]: DEBUG nova.network.neutron [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 669.925674] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff61e242-9258-4890-967e-4aca698341df {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.934155] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f74b7bae-4703-4721-a507-1cda80bc79a0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.963735] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5133972a-e867-4536-a571-d20e22d5af7c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.971502] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d476e19-fd64-4c64-bdab-ed7fe18906e8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 669.984241] env[62753]: DEBUG nova.compute.provider_tree [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 670.000097] env[62753]: DEBUG nova.compute.manager [req-273b4116-f735-448b-88b4-44d898466c51 req-3a13da80-f25f-4da7-8bea-5c97de1e7fb3 service nova] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Received event network-changed-8bf56dd9-d810-4edc-ab37-c57050778d58 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 670.000255] env[62753]: DEBUG nova.compute.manager [req-273b4116-f735-448b-88b4-44d898466c51 req-3a13da80-f25f-4da7-8bea-5c97de1e7fb3 service nova] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Refreshing instance network info cache due to event network-changed-8bf56dd9-d810-4edc-ab37-c57050778d58. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 670.000509] env[62753]: DEBUG oslo_concurrency.lockutils [req-273b4116-f735-448b-88b4-44d898466c51 req-3a13da80-f25f-4da7-8bea-5c97de1e7fb3 service nova] Acquiring lock "refresh_cache-a408799d-29a9-442f-9769-8f8438af8ae0" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 670.181249] env[62753]: DEBUG oslo_concurrency.lockutils [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Releasing lock "refresh_cache-2d028151-3d84-4195-9f93-28287dbfda09" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.181488] env[62753]: DEBUG nova.compute.manager [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 670.181651] env[62753]: DEBUG nova.compute.manager [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 670.181882] env[62753]: DEBUG nova.network.neutron [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 670.198343] env[62753]: DEBUG nova.network.neutron [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.202103] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "refresh_cache-a408799d-29a9-442f-9769-8f8438af8ae0" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 670.202503] env[62753]: DEBUG nova.compute.manager [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 670.202690] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 670.202991] env[62753]: DEBUG oslo_concurrency.lockutils [req-273b4116-f735-448b-88b4-44d898466c51 req-3a13da80-f25f-4da7-8bea-5c97de1e7fb3 service nova] Acquired lock "refresh_cache-a408799d-29a9-442f-9769-8f8438af8ae0" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 670.203176] env[62753]: DEBUG nova.network.neutron [req-273b4116-f735-448b-88b4-44d898466c51 req-3a13da80-f25f-4da7-8bea-5c97de1e7fb3 service nova] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Refreshing network info cache for port 8bf56dd9-d810-4edc-ab37-c57050778d58 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 670.205350] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-778544d8-4ecc-4d00-a90c-9f1d0b90a95e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.215695] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba930365-657b-419a-aaad-f91d10323341 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 670.236840] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a408799d-29a9-442f-9769-8f8438af8ae0 could not be found. [ 670.237071] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 670.237253] env[62753]: INFO nova.compute.manager [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Took 0.03 seconds to destroy the instance on the hypervisor. [ 670.237486] env[62753]: DEBUG oslo.service.loopingcall [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 670.237688] env[62753]: DEBUG nova.compute.manager [-] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 670.237807] env[62753]: DEBUG nova.network.neutron [-] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 670.252233] env[62753]: DEBUG nova.network.neutron [-] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.487485] env[62753]: DEBUG nova.scheduler.client.report [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 670.700558] env[62753]: DEBUG nova.network.neutron [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.723812] env[62753]: DEBUG nova.network.neutron [req-273b4116-f735-448b-88b4-44d898466c51 req-3a13da80-f25f-4da7-8bea-5c97de1e7fb3 service nova] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 670.754675] env[62753]: DEBUG nova.network.neutron [-] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.792683] env[62753]: DEBUG nova.network.neutron [req-273b4116-f735-448b-88b4-44d898466c51 req-3a13da80-f25f-4da7-8bea-5c97de1e7fb3 service nova] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 670.993692] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.475s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 670.994341] env[62753]: ERROR nova.compute.manager [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 05ad2098-191c-4e2e-a2de-94432f0b7a25, please check neutron logs for more information. [ 670.994341] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Traceback (most recent call last): [ 670.994341] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 670.994341] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] self.driver.spawn(context, instance, image_meta, [ 670.994341] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 670.994341] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 670.994341] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 670.994341] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] vm_ref = self.build_virtual_machine(instance, [ 670.994341] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 670.994341] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] vif_infos = vmwarevif.get_vif_info(self._session, [ 670.994341] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 670.994641] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] for vif in network_info: [ 670.994641] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 670.994641] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] return self._sync_wrapper(fn, *args, **kwargs) [ 670.994641] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 670.994641] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] self.wait() [ 670.994641] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 670.994641] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] self[:] = self._gt.wait() [ 670.994641] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 670.994641] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] return self._exit_event.wait() [ 670.994641] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 670.994641] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] result = hub.switch() [ 670.994641] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 670.994641] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] return self.greenlet.switch() [ 670.994926] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 670.994926] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] result = function(*args, **kwargs) [ 670.994926] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 670.994926] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] return func(*args, **kwargs) [ 670.994926] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 670.994926] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] raise e [ 670.994926] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 670.994926] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] nwinfo = self.network_api.allocate_for_instance( [ 670.994926] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 670.994926] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] created_port_ids = self._update_ports_for_instance( [ 670.994926] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 670.994926] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] with excutils.save_and_reraise_exception(): [ 670.994926] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 670.995218] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] self.force_reraise() [ 670.995218] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 670.995218] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] raise self.value [ 670.995218] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 670.995218] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] updated_port = self._update_port( [ 670.995218] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 670.995218] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] _ensure_no_port_binding_failure(port) [ 670.995218] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 670.995218] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] raise exception.PortBindingFailed(port_id=port['id']) [ 670.995218] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] nova.exception.PortBindingFailed: Binding failed for port 05ad2098-191c-4e2e-a2de-94432f0b7a25, please check neutron logs for more information. [ 670.995218] env[62753]: ERROR nova.compute.manager [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] [ 670.995465] env[62753]: DEBUG nova.compute.utils [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Binding failed for port 05ad2098-191c-4e2e-a2de-94432f0b7a25, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 670.996631] env[62753]: DEBUG oslo_concurrency.lockutils [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.286s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 670.999581] env[62753]: DEBUG nova.compute.manager [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Build of instance 989db77e-169f-4b3b-a0b9-c4417f4e89c4 was re-scheduled: Binding failed for port 05ad2098-191c-4e2e-a2de-94432f0b7a25, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 670.999948] env[62753]: DEBUG nova.compute.manager [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 671.000263] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Acquiring lock "refresh_cache-989db77e-169f-4b3b-a0b9-c4417f4e89c4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 671.000417] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Acquired lock "refresh_cache-989db77e-169f-4b3b-a0b9-c4417f4e89c4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 671.000576] env[62753]: DEBUG nova.network.neutron [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 671.203643] env[62753]: INFO nova.compute.manager [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] [instance: 2d028151-3d84-4195-9f93-28287dbfda09] Took 1.02 seconds to deallocate network for instance. [ 671.257240] env[62753]: INFO nova.compute.manager [-] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Took 1.02 seconds to deallocate network for instance. [ 671.259477] env[62753]: DEBUG nova.compute.claims [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 671.259650] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 671.294742] env[62753]: DEBUG oslo_concurrency.lockutils [req-273b4116-f735-448b-88b4-44d898466c51 req-3a13da80-f25f-4da7-8bea-5c97de1e7fb3 service nova] Releasing lock "refresh_cache-a408799d-29a9-442f-9769-8f8438af8ae0" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 671.294870] env[62753]: DEBUG nova.compute.manager [req-273b4116-f735-448b-88b4-44d898466c51 req-3a13da80-f25f-4da7-8bea-5c97de1e7fb3 service nova] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Received event network-vif-deleted-8bf56dd9-d810-4edc-ab37-c57050778d58 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 671.519102] env[62753]: DEBUG nova.network.neutron [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 671.586653] env[62753]: DEBUG nova.network.neutron [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 671.826455] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e7cb684-e3ba-4ffd-b0bf-7811a778ebfd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.834650] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6af96ed8-b3bf-409e-9b6b-c24888b18139 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.865932] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-231e3ecb-f003-406f-9e47-bb7cddc82ddb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.872983] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6850a888-d288-46db-a73e-784f3ed4087b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 671.886238] env[62753]: DEBUG nova.compute.provider_tree [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 672.090749] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Releasing lock "refresh_cache-989db77e-169f-4b3b-a0b9-c4417f4e89c4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 672.091220] env[62753]: DEBUG nova.compute.manager [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 672.091220] env[62753]: DEBUG nova.compute.manager [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 672.091347] env[62753]: DEBUG nova.network.neutron [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 672.107950] env[62753]: DEBUG nova.network.neutron [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 672.232130] env[62753]: INFO nova.scheduler.client.report [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Deleted allocations for instance 2d028151-3d84-4195-9f93-28287dbfda09 [ 672.397367] env[62753]: DEBUG nova.scheduler.client.report [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 672.610689] env[62753]: DEBUG nova.network.neutron [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 672.740460] env[62753]: DEBUG oslo_concurrency.lockutils [None req-276aa15a-8a8f-4190-a20a-6ef4e79c49ce tempest-DeleteServersAdminTestJSON-893351961 tempest-DeleteServersAdminTestJSON-893351961-project-member] Lock "2d028151-3d84-4195-9f93-28287dbfda09" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 76.499s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.903924] env[62753]: DEBUG oslo_concurrency.lockutils [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.907s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 672.904349] env[62753]: ERROR nova.compute.manager [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 43989f12-9ac8-484f-a0e9-ad981ecf1fca, please check neutron logs for more information. [ 672.904349] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Traceback (most recent call last): [ 672.904349] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 672.904349] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] self.driver.spawn(context, instance, image_meta, [ 672.904349] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 672.904349] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] self._vmops.spawn(context, instance, image_meta, injected_files, [ 672.904349] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 672.904349] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] vm_ref = self.build_virtual_machine(instance, [ 672.904349] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 672.904349] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] vif_infos = vmwarevif.get_vif_info(self._session, [ 672.904349] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 672.904705] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] for vif in network_info: [ 672.904705] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 672.904705] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] return self._sync_wrapper(fn, *args, **kwargs) [ 672.904705] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 672.904705] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] self.wait() [ 672.904705] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 672.904705] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] self[:] = self._gt.wait() [ 672.904705] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 672.904705] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] return self._exit_event.wait() [ 672.904705] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 672.904705] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] current.throw(*self._exc) [ 672.904705] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 672.904705] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] result = function(*args, **kwargs) [ 672.905081] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 672.905081] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] return func(*args, **kwargs) [ 672.905081] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 672.905081] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] raise e [ 672.905081] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 672.905081] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] nwinfo = self.network_api.allocate_for_instance( [ 672.905081] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 672.905081] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] created_port_ids = self._update_ports_for_instance( [ 672.905081] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 672.905081] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] with excutils.save_and_reraise_exception(): [ 672.905081] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 672.905081] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] self.force_reraise() [ 672.905081] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 672.905437] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] raise self.value [ 672.905437] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 672.905437] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] updated_port = self._update_port( [ 672.905437] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 672.905437] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] _ensure_no_port_binding_failure(port) [ 672.905437] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 672.905437] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] raise exception.PortBindingFailed(port_id=port['id']) [ 672.905437] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] nova.exception.PortBindingFailed: Binding failed for port 43989f12-9ac8-484f-a0e9-ad981ecf1fca, please check neutron logs for more information. [ 672.905437] env[62753]: ERROR nova.compute.manager [instance: 3c8e531e-033c-4857-9865-b1024e143d44] [ 672.905437] env[62753]: DEBUG nova.compute.utils [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Binding failed for port 43989f12-9ac8-484f-a0e9-ad981ecf1fca, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 672.906672] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.648s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 672.908097] env[62753]: INFO nova.compute.claims [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 672.910807] env[62753]: DEBUG nova.compute.manager [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Build of instance 3c8e531e-033c-4857-9865-b1024e143d44 was re-scheduled: Binding failed for port 43989f12-9ac8-484f-a0e9-ad981ecf1fca, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 672.911340] env[62753]: DEBUG nova.compute.manager [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 672.911891] env[62753]: DEBUG oslo_concurrency.lockutils [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Acquiring lock "refresh_cache-3c8e531e-033c-4857-9865-b1024e143d44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 672.911891] env[62753]: DEBUG oslo_concurrency.lockutils [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Acquired lock "refresh_cache-3c8e531e-033c-4857-9865-b1024e143d44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 672.911979] env[62753]: DEBUG nova.network.neutron [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 673.114057] env[62753]: INFO nova.compute.manager [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] [instance: 989db77e-169f-4b3b-a0b9-c4417f4e89c4] Took 1.02 seconds to deallocate network for instance. [ 673.243519] env[62753]: DEBUG nova.compute.manager [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 673.432097] env[62753]: DEBUG nova.network.neutron [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 673.479513] env[62753]: DEBUG nova.network.neutron [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 673.765051] env[62753]: DEBUG oslo_concurrency.lockutils [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 673.983817] env[62753]: DEBUG oslo_concurrency.lockutils [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Releasing lock "refresh_cache-3c8e531e-033c-4857-9865-b1024e143d44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 673.984087] env[62753]: DEBUG nova.compute.manager [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 673.984224] env[62753]: DEBUG nova.compute.manager [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 673.984390] env[62753]: DEBUG nova.network.neutron [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 674.008153] env[62753]: DEBUG nova.network.neutron [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 674.142631] env[62753]: INFO nova.scheduler.client.report [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Deleted allocations for instance 989db77e-169f-4b3b-a0b9-c4417f4e89c4 [ 674.270949] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aae94c5-9106-4e5a-bf8c-a72ee3ab7e2d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.278756] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a4fe15-03d1-4973-bc45-04b594bb3fa4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.310944] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bdf479b6-fe64-4deb-a08b-a905f6253129 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.319029] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-100411e8-ad57-407f-92d7-dc60c2d47c48 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 674.332482] env[62753]: DEBUG nova.compute.provider_tree [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 674.509558] env[62753]: DEBUG nova.network.neutron [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 674.652424] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea518a74-6953-44b9-8463-c5e5b01ed5c5 tempest-ServersWithSpecificFlavorTestJSON-1756147249 tempest-ServersWithSpecificFlavorTestJSON-1756147249-project-member] Lock "989db77e-169f-4b3b-a0b9-c4417f4e89c4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 77.609s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 674.835373] env[62753]: DEBUG nova.scheduler.client.report [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 675.011946] env[62753]: INFO nova.compute.manager [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] [instance: 3c8e531e-033c-4857-9865-b1024e143d44] Took 1.03 seconds to deallocate network for instance. [ 675.155097] env[62753]: DEBUG nova.compute.manager [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 675.340655] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.434s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.341231] env[62753]: DEBUG nova.compute.manager [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 675.344339] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.126s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.347690] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 675.347690] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62753) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 675.347690] env[62753]: DEBUG oslo_concurrency.lockutils [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.653s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 675.348253] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ed489d-2550-45fc-919e-9694cf3d7c57 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.357105] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4e1df7b-108b-4dca-a915-e0f0d34aa838 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.371254] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d30e15e2-8f56-4fa0-b78c-cf2f350c35a8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.377642] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4258e01-a5d5-4f5b-8365-bcbbe4ee8d93 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 675.409144] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181291MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=62753) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 675.409311] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.679615] env[62753]: DEBUG oslo_concurrency.lockutils [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 675.849565] env[62753]: DEBUG nova.compute.utils [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 675.850969] env[62753]: DEBUG nova.compute.manager [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 675.851157] env[62753]: DEBUG nova.network.neutron [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 675.929736] env[62753]: DEBUG nova.policy [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e0a5dfb7530748168dc5a81bf15fe7d0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '88c5d08819194fc98bd930eaea37fc51', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 676.040968] env[62753]: INFO nova.scheduler.client.report [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Deleted allocations for instance 3c8e531e-033c-4857-9865-b1024e143d44 [ 676.265299] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03b4b4f7-ad78-4618-9fab-83192f84e00e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.273559] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61c089a9-61d2-4900-9003-dae20a1ee519 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.305986] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7397bdb1-4a5f-4172-aa5e-14ab653f4a1f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.313860] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcc24b54-169a-4f28-917b-063201654794 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 676.331295] env[62753]: DEBUG nova.compute.provider_tree [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 676.354041] env[62753]: DEBUG nova.compute.manager [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 676.396289] env[62753]: DEBUG nova.network.neutron [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Successfully created port: 655873cb-9429-4d6c-8a0a-27b6cfabd148 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 676.556415] env[62753]: DEBUG oslo_concurrency.lockutils [None req-52eca1d9-bf25-4ace-809c-7abd8f0c42ce tempest-ServersAdminNegativeTestJSON-1299627216 tempest-ServersAdminNegativeTestJSON-1299627216-project-member] Lock "3c8e531e-033c-4857-9865-b1024e143d44" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 79.356s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 676.837296] env[62753]: DEBUG nova.scheduler.client.report [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 677.061382] env[62753]: DEBUG nova.compute.manager [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 677.265358] env[62753]: DEBUG nova.compute.manager [req-f43343cb-b56d-432a-a27d-b13bcb9d8bad req-3a810376-1fd6-454c-bbe7-3fa7cd2de1e3 service nova] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Received event network-changed-655873cb-9429-4d6c-8a0a-27b6cfabd148 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 677.265358] env[62753]: DEBUG nova.compute.manager [req-f43343cb-b56d-432a-a27d-b13bcb9d8bad req-3a810376-1fd6-454c-bbe7-3fa7cd2de1e3 service nova] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Refreshing instance network info cache due to event network-changed-655873cb-9429-4d6c-8a0a-27b6cfabd148. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 677.265358] env[62753]: DEBUG oslo_concurrency.lockutils [req-f43343cb-b56d-432a-a27d-b13bcb9d8bad req-3a810376-1fd6-454c-bbe7-3fa7cd2de1e3 service nova] Acquiring lock "refresh_cache-d81d5b85-fe31-43ec-91f6-7ff3251ff0c7" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.265358] env[62753]: DEBUG oslo_concurrency.lockutils [req-f43343cb-b56d-432a-a27d-b13bcb9d8bad req-3a810376-1fd6-454c-bbe7-3fa7cd2de1e3 service nova] Acquired lock "refresh_cache-d81d5b85-fe31-43ec-91f6-7ff3251ff0c7" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.265358] env[62753]: DEBUG nova.network.neutron [req-f43343cb-b56d-432a-a27d-b13bcb9d8bad req-3a810376-1fd6-454c-bbe7-3fa7cd2de1e3 service nova] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Refreshing network info cache for port 655873cb-9429-4d6c-8a0a-27b6cfabd148 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 677.343240] env[62753]: DEBUG oslo_concurrency.lockutils [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.995s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 677.343240] env[62753]: ERROR nova.compute.manager [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 13eace83-d89c-4c35-aa6d-541fc4278c56, please check neutron logs for more information. [ 677.343240] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Traceback (most recent call last): [ 677.343240] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.343240] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] self.driver.spawn(context, instance, image_meta, [ 677.343240] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 677.343240] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.343240] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.343240] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] vm_ref = self.build_virtual_machine(instance, [ 677.343645] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.343645] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.343645] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.343645] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] for vif in network_info: [ 677.343645] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.343645] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] return self._sync_wrapper(fn, *args, **kwargs) [ 677.343645] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.343645] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] self.wait() [ 677.343645] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.343645] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] self[:] = self._gt.wait() [ 677.343645] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.343645] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] return self._exit_event.wait() [ 677.343645] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 677.344016] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] result = hub.switch() [ 677.344016] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 677.344016] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] return self.greenlet.switch() [ 677.344016] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.344016] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] result = function(*args, **kwargs) [ 677.344016] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 677.344016] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] return func(*args, **kwargs) [ 677.344016] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.344016] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] raise e [ 677.344016] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.344016] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] nwinfo = self.network_api.allocate_for_instance( [ 677.344016] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 677.344016] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] created_port_ids = self._update_ports_for_instance( [ 677.344311] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 677.344311] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] with excutils.save_and_reraise_exception(): [ 677.344311] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.344311] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] self.force_reraise() [ 677.344311] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.344311] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] raise self.value [ 677.344311] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 677.344311] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] updated_port = self._update_port( [ 677.344311] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.344311] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] _ensure_no_port_binding_failure(port) [ 677.344311] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.344311] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] raise exception.PortBindingFailed(port_id=port['id']) [ 677.344606] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] nova.exception.PortBindingFailed: Binding failed for port 13eace83-d89c-4c35-aa6d-541fc4278c56, please check neutron logs for more information. [ 677.344606] env[62753]: ERROR nova.compute.manager [instance: e730e4f0-fe02-4926-8263-93c8a3065962] [ 677.344606] env[62753]: DEBUG nova.compute.utils [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Binding failed for port 13eace83-d89c-4c35-aa6d-541fc4278c56, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 677.344606] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.983s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 677.349120] env[62753]: DEBUG nova.compute.manager [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Build of instance e730e4f0-fe02-4926-8263-93c8a3065962 was re-scheduled: Binding failed for port 13eace83-d89c-4c35-aa6d-541fc4278c56, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 677.349495] env[62753]: DEBUG nova.compute.manager [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 677.349827] env[62753]: DEBUG oslo_concurrency.lockutils [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Acquiring lock "refresh_cache-e730e4f0-fe02-4926-8263-93c8a3065962" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.350103] env[62753]: DEBUG oslo_concurrency.lockutils [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Acquired lock "refresh_cache-e730e4f0-fe02-4926-8263-93c8a3065962" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 677.350374] env[62753]: DEBUG nova.network.neutron [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 677.365457] env[62753]: DEBUG nova.compute.manager [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 677.407046] env[62753]: DEBUG nova.virt.hardware [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 677.407302] env[62753]: DEBUG nova.virt.hardware [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 677.407870] env[62753]: DEBUG nova.virt.hardware [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 677.407870] env[62753]: DEBUG nova.virt.hardware [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 677.407870] env[62753]: DEBUG nova.virt.hardware [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 677.408012] env[62753]: DEBUG nova.virt.hardware [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 677.408239] env[62753]: DEBUG nova.virt.hardware [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 677.408478] env[62753]: DEBUG nova.virt.hardware [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 677.408687] env[62753]: DEBUG nova.virt.hardware [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 677.408884] env[62753]: DEBUG nova.virt.hardware [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 677.409114] env[62753]: DEBUG nova.virt.hardware [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 677.410857] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e9e3478-8e86-4d56-b476-09fad4fcbe6c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.419862] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dfc7467-b7a1-4892-b777-a991b47fb9b0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 677.527236] env[62753]: ERROR nova.compute.manager [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 655873cb-9429-4d6c-8a0a-27b6cfabd148, please check neutron logs for more information. [ 677.527236] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 677.527236] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.527236] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 677.527236] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 677.527236] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 677.527236] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 677.527236] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 677.527236] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.527236] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 677.527236] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.527236] env[62753]: ERROR nova.compute.manager raise self.value [ 677.527236] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 677.527236] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 677.527236] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.527236] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 677.527658] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.527658] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 677.527658] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 655873cb-9429-4d6c-8a0a-27b6cfabd148, please check neutron logs for more information. [ 677.527658] env[62753]: ERROR nova.compute.manager [ 677.527658] env[62753]: Traceback (most recent call last): [ 677.527658] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 677.527658] env[62753]: listener.cb(fileno) [ 677.527658] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.527658] env[62753]: result = function(*args, **kwargs) [ 677.527658] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 677.527658] env[62753]: return func(*args, **kwargs) [ 677.527658] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.527658] env[62753]: raise e [ 677.527658] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.527658] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 677.527658] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 677.527658] env[62753]: created_port_ids = self._update_ports_for_instance( [ 677.527658] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 677.527658] env[62753]: with excutils.save_and_reraise_exception(): [ 677.527658] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.527658] env[62753]: self.force_reraise() [ 677.527658] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.527658] env[62753]: raise self.value [ 677.527658] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 677.527658] env[62753]: updated_port = self._update_port( [ 677.527658] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.527658] env[62753]: _ensure_no_port_binding_failure(port) [ 677.527658] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.527658] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 677.528377] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 655873cb-9429-4d6c-8a0a-27b6cfabd148, please check neutron logs for more information. [ 677.528377] env[62753]: Removing descriptor: 16 [ 677.528377] env[62753]: ERROR nova.compute.manager [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 655873cb-9429-4d6c-8a0a-27b6cfabd148, please check neutron logs for more information. [ 677.528377] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Traceback (most recent call last): [ 677.528377] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 677.528377] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] yield resources [ 677.528377] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 677.528377] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] self.driver.spawn(context, instance, image_meta, [ 677.528377] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 677.528377] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 677.528377] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 677.528377] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] vm_ref = self.build_virtual_machine(instance, [ 677.528693] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 677.528693] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] vif_infos = vmwarevif.get_vif_info(self._session, [ 677.528693] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 677.528693] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] for vif in network_info: [ 677.528693] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 677.528693] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] return self._sync_wrapper(fn, *args, **kwargs) [ 677.528693] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 677.528693] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] self.wait() [ 677.528693] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 677.528693] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] self[:] = self._gt.wait() [ 677.528693] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 677.528693] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] return self._exit_event.wait() [ 677.528693] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 677.529025] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] result = hub.switch() [ 677.529025] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 677.529025] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] return self.greenlet.switch() [ 677.529025] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 677.529025] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] result = function(*args, **kwargs) [ 677.529025] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 677.529025] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] return func(*args, **kwargs) [ 677.529025] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 677.529025] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] raise e [ 677.529025] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 677.529025] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] nwinfo = self.network_api.allocate_for_instance( [ 677.529025] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 677.529025] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] created_port_ids = self._update_ports_for_instance( [ 677.529359] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 677.529359] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] with excutils.save_and_reraise_exception(): [ 677.529359] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 677.529359] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] self.force_reraise() [ 677.529359] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 677.529359] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] raise self.value [ 677.529359] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 677.529359] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] updated_port = self._update_port( [ 677.529359] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 677.529359] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] _ensure_no_port_binding_failure(port) [ 677.529359] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 677.529359] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] raise exception.PortBindingFailed(port_id=port['id']) [ 677.529657] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] nova.exception.PortBindingFailed: Binding failed for port 655873cb-9429-4d6c-8a0a-27b6cfabd148, please check neutron logs for more information. [ 677.529657] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] [ 677.529657] env[62753]: INFO nova.compute.manager [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Terminating instance [ 677.534213] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Acquiring lock "refresh_cache-d81d5b85-fe31-43ec-91f6-7ff3251ff0c7" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 677.586794] env[62753]: DEBUG oslo_concurrency.lockutils [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 677.880139] env[62753]: DEBUG nova.network.neutron [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 677.931021] env[62753]: DEBUG nova.network.neutron [req-f43343cb-b56d-432a-a27d-b13bcb9d8bad req-3a810376-1fd6-454c-bbe7-3fa7cd2de1e3 service nova] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.014216] env[62753]: DEBUG nova.network.neutron [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.172027] env[62753]: DEBUG nova.network.neutron [req-f43343cb-b56d-432a-a27d-b13bcb9d8bad req-3a810376-1fd6-454c-bbe7-3fa7cd2de1e3 service nova] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 678.306123] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ac7b17a-4e38-469a-91a2-87b9bca1cc38 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.316115] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a38bb2e-090d-451f-9a5d-20c40fe91f6f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.347468] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a38f4f54-644d-4c4c-86f5-c072a7d88da6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.355361] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19938621-ca00-4a53-8750-ffb82940e3c3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 678.370409] env[62753]: DEBUG nova.compute.provider_tree [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 678.520080] env[62753]: DEBUG oslo_concurrency.lockutils [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Releasing lock "refresh_cache-e730e4f0-fe02-4926-8263-93c8a3065962" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.520332] env[62753]: DEBUG nova.compute.manager [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 678.520516] env[62753]: DEBUG nova.compute.manager [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 678.520690] env[62753]: DEBUG nova.network.neutron [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 678.545575] env[62753]: DEBUG nova.network.neutron [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 678.676862] env[62753]: DEBUG oslo_concurrency.lockutils [req-f43343cb-b56d-432a-a27d-b13bcb9d8bad req-3a810376-1fd6-454c-bbe7-3fa7cd2de1e3 service nova] Releasing lock "refresh_cache-d81d5b85-fe31-43ec-91f6-7ff3251ff0c7" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 678.677383] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Acquired lock "refresh_cache-d81d5b85-fe31-43ec-91f6-7ff3251ff0c7" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 678.677589] env[62753]: DEBUG nova.network.neutron [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 678.878294] env[62753]: DEBUG nova.scheduler.client.report [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 679.049268] env[62753]: DEBUG nova.network.neutron [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.201248] env[62753]: DEBUG nova.network.neutron [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.335081] env[62753]: DEBUG nova.network.neutron [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 679.352292] env[62753]: DEBUG nova.compute.manager [req-d9144c29-eced-4b6c-8268-a1f01fa867fc req-5e14b79e-2232-42d8-abf8-e4ea4934731f service nova] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Received event network-vif-deleted-655873cb-9429-4d6c-8a0a-27b6cfabd148 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 679.386808] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.044s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 679.387932] env[62753]: ERROR nova.compute.manager [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ce5b0145-14e4-46d5-90da-42f0588ecc04, please check neutron logs for more information. [ 679.387932] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Traceback (most recent call last): [ 679.387932] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 679.387932] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] self.driver.spawn(context, instance, image_meta, [ 679.387932] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 679.387932] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] self._vmops.spawn(context, instance, image_meta, injected_files, [ 679.387932] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 679.387932] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] vm_ref = self.build_virtual_machine(instance, [ 679.387932] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 679.387932] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] vif_infos = vmwarevif.get_vif_info(self._session, [ 679.387932] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 679.388480] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] for vif in network_info: [ 679.388480] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 679.388480] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] return self._sync_wrapper(fn, *args, **kwargs) [ 679.388480] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 679.388480] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] self.wait() [ 679.388480] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 679.388480] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] self[:] = self._gt.wait() [ 679.388480] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 679.388480] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] return self._exit_event.wait() [ 679.388480] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 679.388480] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] current.throw(*self._exc) [ 679.388480] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 679.388480] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] result = function(*args, **kwargs) [ 679.388883] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 679.388883] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] return func(*args, **kwargs) [ 679.388883] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 679.388883] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] raise e [ 679.388883] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 679.388883] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] nwinfo = self.network_api.allocate_for_instance( [ 679.388883] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 679.388883] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] created_port_ids = self._update_ports_for_instance( [ 679.388883] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 679.388883] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] with excutils.save_and_reraise_exception(): [ 679.388883] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 679.388883] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] self.force_reraise() [ 679.388883] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 679.389314] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] raise self.value [ 679.389314] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 679.389314] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] updated_port = self._update_port( [ 679.389314] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 679.389314] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] _ensure_no_port_binding_failure(port) [ 679.389314] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 679.389314] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] raise exception.PortBindingFailed(port_id=port['id']) [ 679.389314] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] nova.exception.PortBindingFailed: Binding failed for port ce5b0145-14e4-46d5-90da-42f0588ecc04, please check neutron logs for more information. [ 679.389314] env[62753]: ERROR nova.compute.manager [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] [ 679.389691] env[62753]: DEBUG nova.compute.utils [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Binding failed for port ce5b0145-14e4-46d5-90da-42f0588ecc04, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 679.394910] env[62753]: DEBUG oslo_concurrency.lockutils [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.538s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 679.397517] env[62753]: INFO nova.compute.claims [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 679.401048] env[62753]: DEBUG nova.compute.manager [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Build of instance e8f80847-a14d-4abd-af9f-84c62dd1e395 was re-scheduled: Binding failed for port ce5b0145-14e4-46d5-90da-42f0588ecc04, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 679.401336] env[62753]: DEBUG nova.compute.manager [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 679.401563] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Acquiring lock "refresh_cache-e8f80847-a14d-4abd-af9f-84c62dd1e395" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 679.401709] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Acquired lock "refresh_cache-e8f80847-a14d-4abd-af9f-84c62dd1e395" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 679.404808] env[62753]: DEBUG nova.network.neutron [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 679.555740] env[62753]: INFO nova.compute.manager [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] [instance: e730e4f0-fe02-4926-8263-93c8a3065962] Took 1.03 seconds to deallocate network for instance. [ 679.842377] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Releasing lock "refresh_cache-d81d5b85-fe31-43ec-91f6-7ff3251ff0c7" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 679.842377] env[62753]: DEBUG nova.compute.manager [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 679.842377] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 679.842377] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0bb4dbc7-00cf-47d6-b70f-a5ee090725f9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.852730] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3953e3c-464e-4f85-a321-57ff8e4ff8bc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 679.881275] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d81d5b85-fe31-43ec-91f6-7ff3251ff0c7 could not be found. [ 679.881546] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 679.881728] env[62753]: INFO nova.compute.manager [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Took 0.04 seconds to destroy the instance on the hypervisor. [ 679.883102] env[62753]: DEBUG oslo.service.loopingcall [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 679.883102] env[62753]: DEBUG nova.compute.manager [-] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 679.883102] env[62753]: DEBUG nova.network.neutron [-] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 679.897952] env[62753]: DEBUG nova.network.neutron [-] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 679.920080] env[62753]: DEBUG nova.network.neutron [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.005848] env[62753]: DEBUG nova.network.neutron [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.329416] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Acquiring lock "e0c4a1d2-6a85-4826-88fa-d21556d744ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.329698] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Lock "e0c4a1d2-6a85-4826-88fa-d21556d744ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 680.403670] env[62753]: DEBUG nova.network.neutron [-] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 680.510593] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Releasing lock "refresh_cache-e8f80847-a14d-4abd-af9f-84c62dd1e395" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 680.510593] env[62753]: DEBUG nova.compute.manager [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 680.510593] env[62753]: DEBUG nova.compute.manager [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 680.510593] env[62753]: DEBUG nova.network.neutron [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 680.529219] env[62753]: DEBUG nova.network.neutron [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 680.591714] env[62753]: INFO nova.scheduler.client.report [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Deleted allocations for instance e730e4f0-fe02-4926-8263-93c8a3065962 [ 680.857508] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c78b4ba-34d5-489f-a203-a5f6869e3da5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.866685] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae3ad59e-e0a9-417f-9e6f-007e4d8284b4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.900440] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00db76d3-5cd0-4466-b36b-a5d5456f468e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.908944] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc469c78-5144-4873-ba0d-dd73d3a16e3d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 680.913226] env[62753]: INFO nova.compute.manager [-] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Took 1.03 seconds to deallocate network for instance. [ 680.915307] env[62753]: DEBUG nova.compute.claims [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 680.915443] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 680.923975] env[62753]: DEBUG nova.compute.provider_tree [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 681.038128] env[62753]: DEBUG nova.network.neutron [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 681.107397] env[62753]: DEBUG oslo_concurrency.lockutils [None req-19ac8f09-be7b-406d-8073-ac94e9e7a70e tempest-AttachInterfacesUnderV243Test-2056480492 tempest-AttachInterfacesUnderV243Test-2056480492-project-member] Lock "e730e4f0-fe02-4926-8263-93c8a3065962" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.127s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.427734] env[62753]: DEBUG nova.scheduler.client.report [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 681.541757] env[62753]: INFO nova.compute.manager [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] [instance: e8f80847-a14d-4abd-af9f-84c62dd1e395] Took 1.03 seconds to deallocate network for instance. [ 681.611959] env[62753]: DEBUG nova.compute.manager [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 681.933718] env[62753]: DEBUG oslo_concurrency.lockutils [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.542s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 681.934669] env[62753]: DEBUG nova.compute.manager [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 681.937903] env[62753]: DEBUG oslo_concurrency.lockutils [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.889s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 682.141263] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 682.448289] env[62753]: DEBUG nova.compute.utils [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 682.455568] env[62753]: DEBUG nova.compute.manager [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 682.455568] env[62753]: DEBUG nova.network.neutron [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 682.579118] env[62753]: INFO nova.scheduler.client.report [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Deleted allocations for instance e8f80847-a14d-4abd-af9f-84c62dd1e395 [ 682.661889] env[62753]: DEBUG nova.policy [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '613b397bebf9498596d07d27b9fa1fa6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b16df868b31c4f6597972c6f2efa9494', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 682.883715] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15a5be86-045b-46b0-8f62-bbfd2d4114df {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.891664] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f945455-252b-4b39-93c0-e63309b97766 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.927853] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9aa6ce81-6e67-43b9-9c5d-4ad6c3adb255 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.935556] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbada775-3570-477b-ac01-f3eb66cdd44a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 682.950927] env[62753]: DEBUG nova.compute.provider_tree [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 682.953638] env[62753]: DEBUG nova.compute.manager [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 683.092521] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b5e2e0d-20e2-4c03-a185-7f2c8e3fec06 tempest-FloatingIPsAssociationTestJSON-683248272 tempest-FloatingIPsAssociationTestJSON-683248272-project-member] Lock "e8f80847-a14d-4abd-af9f-84c62dd1e395" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.437s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.162419] env[62753]: DEBUG nova.network.neutron [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Successfully created port: d93e8619-1348-4566-9c18-ebbebb6b1f46 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 683.277532] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquiring lock "41476adb-95e5-4617-b178-9c981d13c43d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.277532] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Lock "41476adb-95e5-4617-b178-9c981d13c43d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.458897] env[62753]: DEBUG nova.scheduler.client.report [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 683.596498] env[62753]: DEBUG nova.compute.manager [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 683.758491] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Acquiring lock "3adb208d-cf70-4d93-b034-a3e87d029f90" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 683.758717] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Lock "3adb208d-cf70-4d93-b034-a3e87d029f90" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 683.974252] env[62753]: DEBUG oslo_concurrency.lockutils [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.036s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 683.975680] env[62753]: ERROR nova.compute.manager [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ad3348fb-2ed5-4d71-adfb-edfbeb299328, please check neutron logs for more information. [ 683.975680] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Traceback (most recent call last): [ 683.975680] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 683.975680] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] self.driver.spawn(context, instance, image_meta, [ 683.975680] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 683.975680] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 683.975680] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 683.975680] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] vm_ref = self.build_virtual_machine(instance, [ 683.975680] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 683.975680] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] vif_infos = vmwarevif.get_vif_info(self._session, [ 683.975680] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 683.975990] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] for vif in network_info: [ 683.975990] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 683.975990] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] return self._sync_wrapper(fn, *args, **kwargs) [ 683.975990] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 683.975990] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] self.wait() [ 683.975990] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 683.975990] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] self[:] = self._gt.wait() [ 683.975990] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 683.975990] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] return self._exit_event.wait() [ 683.975990] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 683.975990] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] result = hub.switch() [ 683.975990] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 683.975990] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] return self.greenlet.switch() [ 683.976324] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 683.976324] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] result = function(*args, **kwargs) [ 683.976324] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 683.976324] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] return func(*args, **kwargs) [ 683.976324] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 683.976324] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] raise e [ 683.976324] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 683.976324] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] nwinfo = self.network_api.allocate_for_instance( [ 683.976324] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 683.976324] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] created_port_ids = self._update_ports_for_instance( [ 683.976324] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 683.976324] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] with excutils.save_and_reraise_exception(): [ 683.976324] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 683.976645] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] self.force_reraise() [ 683.976645] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 683.976645] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] raise self.value [ 683.976645] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 683.976645] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] updated_port = self._update_port( [ 683.976645] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 683.976645] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] _ensure_no_port_binding_failure(port) [ 683.976645] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 683.976645] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] raise exception.PortBindingFailed(port_id=port['id']) [ 683.976645] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] nova.exception.PortBindingFailed: Binding failed for port ad3348fb-2ed5-4d71-adfb-edfbeb299328, please check neutron logs for more information. [ 683.976645] env[62753]: ERROR nova.compute.manager [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] [ 683.978052] env[62753]: DEBUG nova.compute.utils [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Binding failed for port ad3348fb-2ed5-4d71-adfb-edfbeb299328, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 683.983091] env[62753]: DEBUG nova.compute.manager [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 683.985658] env[62753]: DEBUG nova.compute.manager [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Build of instance 3d13009d-5e65-43e1-87b3-3cb3cfd014e6 was re-scheduled: Binding failed for port ad3348fb-2ed5-4d71-adfb-edfbeb299328, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 683.986538] env[62753]: DEBUG nova.compute.manager [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 683.986836] env[62753]: DEBUG oslo_concurrency.lockutils [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Acquiring lock "refresh_cache-3d13009d-5e65-43e1-87b3-3cb3cfd014e6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 683.986999] env[62753]: DEBUG oslo_concurrency.lockutils [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Acquired lock "refresh_cache-3d13009d-5e65-43e1-87b3-3cb3cfd014e6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 683.987207] env[62753]: DEBUG nova.network.neutron [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 683.988215] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.384s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 684.016023] env[62753]: DEBUG nova.virt.hardware [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 684.016257] env[62753]: DEBUG nova.virt.hardware [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 684.016408] env[62753]: DEBUG nova.virt.hardware [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 684.016586] env[62753]: DEBUG nova.virt.hardware [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 684.016725] env[62753]: DEBUG nova.virt.hardware [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 684.016871] env[62753]: DEBUG nova.virt.hardware [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 684.019310] env[62753]: DEBUG nova.virt.hardware [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 684.019310] env[62753]: DEBUG nova.virt.hardware [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 684.019310] env[62753]: DEBUG nova.virt.hardware [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 684.019889] env[62753]: DEBUG nova.virt.hardware [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 684.019889] env[62753]: DEBUG nova.virt.hardware [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 684.020961] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa7d99f0-8702-426d-aa3c-4da49045ba3a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.030372] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c09e118-f212-4152-9b2d-6262c3f00b09 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.122650] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 684.280045] env[62753]: DEBUG nova.compute.manager [req-12c3fc0a-bf83-415d-a8fe-b553de7e05ad req-3b013ea5-940f-4eef-a7c3-669914ffd210 service nova] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Received event network-changed-d93e8619-1348-4566-9c18-ebbebb6b1f46 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 684.280251] env[62753]: DEBUG nova.compute.manager [req-12c3fc0a-bf83-415d-a8fe-b553de7e05ad req-3b013ea5-940f-4eef-a7c3-669914ffd210 service nova] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Refreshing instance network info cache due to event network-changed-d93e8619-1348-4566-9c18-ebbebb6b1f46. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 684.280464] env[62753]: DEBUG oslo_concurrency.lockutils [req-12c3fc0a-bf83-415d-a8fe-b553de7e05ad req-3b013ea5-940f-4eef-a7c3-669914ffd210 service nova] Acquiring lock "refresh_cache-f10ff402-4308-4c75-af79-26e2942e0b92" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.280607] env[62753]: DEBUG oslo_concurrency.lockutils [req-12c3fc0a-bf83-415d-a8fe-b553de7e05ad req-3b013ea5-940f-4eef-a7c3-669914ffd210 service nova] Acquired lock "refresh_cache-f10ff402-4308-4c75-af79-26e2942e0b92" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 684.280772] env[62753]: DEBUG nova.network.neutron [req-12c3fc0a-bf83-415d-a8fe-b553de7e05ad req-3b013ea5-940f-4eef-a7c3-669914ffd210 service nova] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Refreshing network info cache for port d93e8619-1348-4566-9c18-ebbebb6b1f46 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 684.522939] env[62753]: DEBUG nova.network.neutron [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.566530] env[62753]: ERROR nova.compute.manager [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port d93e8619-1348-4566-9c18-ebbebb6b1f46, please check neutron logs for more information. [ 684.566530] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 684.566530] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.566530] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 684.566530] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.566530] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 684.566530] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.566530] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 684.566530] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.566530] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 684.566530] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.566530] env[62753]: ERROR nova.compute.manager raise self.value [ 684.566530] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.566530] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 684.566530] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.566530] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 684.567014] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.567014] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 684.567014] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port d93e8619-1348-4566-9c18-ebbebb6b1f46, please check neutron logs for more information. [ 684.567014] env[62753]: ERROR nova.compute.manager [ 684.568024] env[62753]: Traceback (most recent call last): [ 684.568024] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 684.568024] env[62753]: listener.cb(fileno) [ 684.568024] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.568024] env[62753]: result = function(*args, **kwargs) [ 684.568024] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 684.568024] env[62753]: return func(*args, **kwargs) [ 684.568024] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.568024] env[62753]: raise e [ 684.568024] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.568024] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 684.568024] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.568024] env[62753]: created_port_ids = self._update_ports_for_instance( [ 684.568024] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.568024] env[62753]: with excutils.save_and_reraise_exception(): [ 684.568024] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.568024] env[62753]: self.force_reraise() [ 684.568024] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.568024] env[62753]: raise self.value [ 684.568024] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.568024] env[62753]: updated_port = self._update_port( [ 684.568024] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.568024] env[62753]: _ensure_no_port_binding_failure(port) [ 684.568024] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.568024] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 684.568024] env[62753]: nova.exception.PortBindingFailed: Binding failed for port d93e8619-1348-4566-9c18-ebbebb6b1f46, please check neutron logs for more information. [ 684.568024] env[62753]: Removing descriptor: 16 [ 684.571158] env[62753]: ERROR nova.compute.manager [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port d93e8619-1348-4566-9c18-ebbebb6b1f46, please check neutron logs for more information. [ 684.571158] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Traceback (most recent call last): [ 684.571158] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 684.571158] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] yield resources [ 684.571158] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 684.571158] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] self.driver.spawn(context, instance, image_meta, [ 684.571158] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 684.571158] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] self._vmops.spawn(context, instance, image_meta, injected_files, [ 684.571158] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 684.571158] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] vm_ref = self.build_virtual_machine(instance, [ 684.571158] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 684.571569] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] vif_infos = vmwarevif.get_vif_info(self._session, [ 684.571569] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 684.571569] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] for vif in network_info: [ 684.571569] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 684.571569] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] return self._sync_wrapper(fn, *args, **kwargs) [ 684.571569] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 684.571569] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] self.wait() [ 684.571569] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 684.571569] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] self[:] = self._gt.wait() [ 684.571569] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 684.571569] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] return self._exit_event.wait() [ 684.571569] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 684.571569] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] result = hub.switch() [ 684.571968] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 684.571968] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] return self.greenlet.switch() [ 684.571968] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 684.571968] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] result = function(*args, **kwargs) [ 684.571968] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 684.571968] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] return func(*args, **kwargs) [ 684.571968] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 684.571968] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] raise e [ 684.571968] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 684.571968] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] nwinfo = self.network_api.allocate_for_instance( [ 684.571968] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 684.571968] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] created_port_ids = self._update_ports_for_instance( [ 684.571968] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 684.572336] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] with excutils.save_and_reraise_exception(): [ 684.572336] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 684.572336] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] self.force_reraise() [ 684.572336] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 684.572336] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] raise self.value [ 684.572336] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 684.572336] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] updated_port = self._update_port( [ 684.572336] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 684.572336] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] _ensure_no_port_binding_failure(port) [ 684.572336] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 684.572336] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] raise exception.PortBindingFailed(port_id=port['id']) [ 684.572336] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] nova.exception.PortBindingFailed: Binding failed for port d93e8619-1348-4566-9c18-ebbebb6b1f46, please check neutron logs for more information. [ 684.572336] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] [ 684.572704] env[62753]: INFO nova.compute.manager [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Terminating instance [ 684.574858] env[62753]: DEBUG oslo_concurrency.lockutils [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Acquiring lock "refresh_cache-f10ff402-4308-4c75-af79-26e2942e0b92" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 684.654034] env[62753]: DEBUG nova.network.neutron [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.806998] env[62753]: DEBUG nova.network.neutron [req-12c3fc0a-bf83-415d-a8fe-b553de7e05ad req-3b013ea5-940f-4eef-a7c3-669914ffd210 service nova] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 684.915196] env[62753]: DEBUG nova.network.neutron [req-12c3fc0a-bf83-415d-a8fe-b553de7e05ad req-3b013ea5-940f-4eef-a7c3-669914ffd210 service nova] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 684.932631] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3faa0b9-0cbf-4e46-a8f2-19f1457ca6b1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.942611] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8403c453-c064-4b8a-aaf6-63b3b49f98f8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.975562] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abf0ea4e-e6e2-4581-8bae-df7e7814b33c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.983666] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ed3514-0880-4448-8ea3-224794934ae9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 684.997790] env[62753]: DEBUG nova.compute.provider_tree [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 685.156289] env[62753]: DEBUG oslo_concurrency.lockutils [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Releasing lock "refresh_cache-3d13009d-5e65-43e1-87b3-3cb3cfd014e6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.156622] env[62753]: DEBUG nova.compute.manager [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 685.156683] env[62753]: DEBUG nova.compute.manager [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 685.156873] env[62753]: DEBUG nova.network.neutron [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 685.176110] env[62753]: DEBUG nova.network.neutron [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 685.418027] env[62753]: DEBUG oslo_concurrency.lockutils [req-12c3fc0a-bf83-415d-a8fe-b553de7e05ad req-3b013ea5-940f-4eef-a7c3-669914ffd210 service nova] Releasing lock "refresh_cache-f10ff402-4308-4c75-af79-26e2942e0b92" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 685.422318] env[62753]: DEBUG oslo_concurrency.lockutils [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Acquired lock "refresh_cache-f10ff402-4308-4c75-af79-26e2942e0b92" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 685.422465] env[62753]: DEBUG nova.network.neutron [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 685.505072] env[62753]: DEBUG nova.scheduler.client.report [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 685.678719] env[62753]: DEBUG nova.network.neutron [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 685.942984] env[62753]: DEBUG nova.network.neutron [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.012269] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.021s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 686.012269] env[62753]: ERROR nova.compute.manager [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c2fd8ca2-0060-40ba-a71d-4804b9f9e9f8, please check neutron logs for more information. [ 686.012269] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Traceback (most recent call last): [ 686.012269] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 686.012269] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] self.driver.spawn(context, instance, image_meta, [ 686.012269] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 686.012269] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 686.012269] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 686.012269] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] vm_ref = self.build_virtual_machine(instance, [ 686.012764] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 686.012764] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] vif_infos = vmwarevif.get_vif_info(self._session, [ 686.012764] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 686.012764] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] for vif in network_info: [ 686.012764] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 686.012764] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] return self._sync_wrapper(fn, *args, **kwargs) [ 686.012764] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 686.012764] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] self.wait() [ 686.012764] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 686.012764] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] self[:] = self._gt.wait() [ 686.012764] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 686.012764] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] return self._exit_event.wait() [ 686.012764] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 686.013140] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] result = hub.switch() [ 686.013140] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 686.013140] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] return self.greenlet.switch() [ 686.013140] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 686.013140] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] result = function(*args, **kwargs) [ 686.013140] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 686.013140] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] return func(*args, **kwargs) [ 686.013140] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 686.013140] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] raise e [ 686.013140] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 686.013140] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] nwinfo = self.network_api.allocate_for_instance( [ 686.013140] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 686.013140] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] created_port_ids = self._update_ports_for_instance( [ 686.013445] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 686.013445] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] with excutils.save_and_reraise_exception(): [ 686.013445] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 686.013445] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] self.force_reraise() [ 686.013445] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 686.013445] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] raise self.value [ 686.013445] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 686.013445] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] updated_port = self._update_port( [ 686.013445] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 686.013445] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] _ensure_no_port_binding_failure(port) [ 686.013445] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 686.013445] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] raise exception.PortBindingFailed(port_id=port['id']) [ 686.013806] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] nova.exception.PortBindingFailed: Binding failed for port c2fd8ca2-0060-40ba-a71d-4804b9f9e9f8, please check neutron logs for more information. [ 686.013806] env[62753]: ERROR nova.compute.manager [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] [ 686.013806] env[62753]: DEBUG nova.compute.utils [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Binding failed for port c2fd8ca2-0060-40ba-a71d-4804b9f9e9f8, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 686.013806] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.752s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 686.023737] env[62753]: DEBUG nova.compute.manager [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Build of instance b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a was re-scheduled: Binding failed for port c2fd8ca2-0060-40ba-a71d-4804b9f9e9f8, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 686.023737] env[62753]: DEBUG nova.compute.manager [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 686.023737] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Acquiring lock "refresh_cache-b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 686.023737] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Acquired lock "refresh_cache-b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 686.024036] env[62753]: DEBUG nova.network.neutron [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 686.111292] env[62753]: DEBUG nova.network.neutron [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.183752] env[62753]: INFO nova.compute.manager [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: 3d13009d-5e65-43e1-87b3-3cb3cfd014e6] Took 1.03 seconds to deallocate network for instance. [ 686.314746] env[62753]: DEBUG nova.compute.manager [req-775075b3-56a0-4afe-bc5b-9d86b56a80a6 req-0af3b658-da79-4782-81e2-eb5dbae4189a service nova] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Received event network-vif-deleted-d93e8619-1348-4566-9c18-ebbebb6b1f46 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 686.564211] env[62753]: DEBUG nova.network.neutron [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.613824] env[62753]: DEBUG oslo_concurrency.lockutils [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Releasing lock "refresh_cache-f10ff402-4308-4c75-af79-26e2942e0b92" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 686.614348] env[62753]: DEBUG nova.compute.manager [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 686.614426] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 686.614707] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d2b8b0e6-b9d5-4714-9c26-5b35d3cb5986 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.625500] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac6b8666-b8ea-49b5-b6ed-82cf1e9f245b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.648203] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f10ff402-4308-4c75-af79-26e2942e0b92 could not be found. [ 686.648479] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 686.648722] env[62753]: INFO nova.compute.manager [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Took 0.03 seconds to destroy the instance on the hypervisor. [ 686.649022] env[62753]: DEBUG oslo.service.loopingcall [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 686.651873] env[62753]: DEBUG nova.compute.manager [-] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 686.652029] env[62753]: DEBUG nova.network.neutron [-] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 686.654121] env[62753]: DEBUG nova.network.neutron [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 686.669998] env[62753]: DEBUG nova.network.neutron [-] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 686.981046] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72cba08f-fdca-4047-b031-61b62b12d57a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 686.989732] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d40613b-f524-4e1d-b4e9-10ba4ce851ee {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.025237] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0eb52a7e-e700-4bd2-ad19-e605dfb2a17d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.034155] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cd21979-37bb-409e-9079-e8395ce6cecd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 687.046658] env[62753]: DEBUG nova.compute.provider_tree [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 687.156935] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Releasing lock "refresh_cache-b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 687.157242] env[62753]: DEBUG nova.compute.manager [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 687.157464] env[62753]: DEBUG nova.compute.manager [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 687.157536] env[62753]: DEBUG nova.network.neutron [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 687.173331] env[62753]: DEBUG nova.network.neutron [-] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.176670] env[62753]: DEBUG nova.network.neutron [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 687.217979] env[62753]: INFO nova.scheduler.client.report [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Deleted allocations for instance 3d13009d-5e65-43e1-87b3-3cb3cfd014e6 [ 687.550726] env[62753]: DEBUG nova.scheduler.client.report [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 687.676867] env[62753]: INFO nova.compute.manager [-] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Took 1.02 seconds to deallocate network for instance. [ 687.679141] env[62753]: DEBUG nova.network.neutron [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 687.680854] env[62753]: DEBUG nova.compute.claims [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 687.681136] env[62753]: DEBUG oslo_concurrency.lockutils [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 687.726537] env[62753]: DEBUG oslo_concurrency.lockutils [None req-81065724-6474-4cd0-9ab0-9a606366ada8 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Lock "3d13009d-5e65-43e1-87b3-3cb3cfd014e6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 88.756s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.056495] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.044s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 688.057151] env[62753]: ERROR nova.compute.manager [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8bf56dd9-d810-4edc-ab37-c57050778d58, please check neutron logs for more information. [ 688.057151] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Traceback (most recent call last): [ 688.057151] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 688.057151] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] self.driver.spawn(context, instance, image_meta, [ 688.057151] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 688.057151] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] self._vmops.spawn(context, instance, image_meta, injected_files, [ 688.057151] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 688.057151] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] vm_ref = self.build_virtual_machine(instance, [ 688.057151] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 688.057151] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] vif_infos = vmwarevif.get_vif_info(self._session, [ 688.057151] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 688.057511] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] for vif in network_info: [ 688.057511] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 688.057511] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] return self._sync_wrapper(fn, *args, **kwargs) [ 688.057511] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 688.057511] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] self.wait() [ 688.057511] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 688.057511] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] self[:] = self._gt.wait() [ 688.057511] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 688.057511] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] return self._exit_event.wait() [ 688.057511] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 688.057511] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] current.throw(*self._exc) [ 688.057511] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 688.057511] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] result = function(*args, **kwargs) [ 688.057867] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 688.057867] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] return func(*args, **kwargs) [ 688.057867] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 688.057867] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] raise e [ 688.057867] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 688.057867] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] nwinfo = self.network_api.allocate_for_instance( [ 688.057867] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 688.057867] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] created_port_ids = self._update_ports_for_instance( [ 688.057867] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 688.057867] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] with excutils.save_and_reraise_exception(): [ 688.057867] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 688.057867] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] self.force_reraise() [ 688.057867] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 688.058177] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] raise self.value [ 688.058177] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 688.058177] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] updated_port = self._update_port( [ 688.058177] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 688.058177] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] _ensure_no_port_binding_failure(port) [ 688.058177] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 688.058177] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] raise exception.PortBindingFailed(port_id=port['id']) [ 688.058177] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] nova.exception.PortBindingFailed: Binding failed for port 8bf56dd9-d810-4edc-ab37-c57050778d58, please check neutron logs for more information. [ 688.058177] env[62753]: ERROR nova.compute.manager [instance: a408799d-29a9-442f-9769-8f8438af8ae0] [ 688.058177] env[62753]: DEBUG nova.compute.utils [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Binding failed for port 8bf56dd9-d810-4edc-ab37-c57050778d58, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 688.059078] env[62753]: DEBUG oslo_concurrency.lockutils [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.294s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 688.060539] env[62753]: INFO nova.compute.claims [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 688.063456] env[62753]: DEBUG nova.compute.manager [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Build of instance a408799d-29a9-442f-9769-8f8438af8ae0 was re-scheduled: Binding failed for port 8bf56dd9-d810-4edc-ab37-c57050778d58, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 688.063888] env[62753]: DEBUG nova.compute.manager [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 688.064168] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "refresh_cache-a408799d-29a9-442f-9769-8f8438af8ae0" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 688.064279] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "refresh_cache-a408799d-29a9-442f-9769-8f8438af8ae0" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 688.064433] env[62753]: DEBUG nova.network.neutron [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 688.182587] env[62753]: INFO nova.compute.manager [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] [instance: b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a] Took 1.02 seconds to deallocate network for instance. [ 688.230188] env[62753]: DEBUG nova.compute.manager [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 688.589654] env[62753]: DEBUG nova.network.neutron [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 688.709264] env[62753]: DEBUG nova.network.neutron [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 688.758229] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.215792] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "refresh_cache-a408799d-29a9-442f-9769-8f8438af8ae0" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 689.216429] env[62753]: DEBUG nova.compute.manager [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 689.216796] env[62753]: DEBUG nova.compute.manager [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 689.217047] env[62753]: DEBUG nova.network.neutron [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 689.223090] env[62753]: INFO nova.scheduler.client.report [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Deleted allocations for instance b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a [ 689.246029] env[62753]: DEBUG nova.network.neutron [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 689.483751] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Acquiring lock "f63a8912-0abf-4c76-aaae-4085b3909688" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 689.483996] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Lock "f63a8912-0abf-4c76-aaae-4085b3909688" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 689.531229] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22f5c3fd-925e-45ae-b2db-a3851c4cf875 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.539870] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-554991e5-1095-49be-abb4-468c10ca7f48 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.575554] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91d0623f-0085-483e-9106-501f3a69d471 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.582966] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-664e0389-261f-4244-ab8f-5a12e047cc5a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 689.598143] env[62753]: DEBUG nova.compute.provider_tree [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 689.740303] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4111ef-e5e1-4daa-8e10-a98751c67039 tempest-TenantUsagesTestJSON-1326507039 tempest-TenantUsagesTestJSON-1326507039-project-member] Lock "b4b4b7d3-d78e-4c6d-986f-42dffacf0c5a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.134s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 689.753435] env[62753]: DEBUG nova.network.neutron [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 690.100751] env[62753]: DEBUG nova.scheduler.client.report [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 690.244930] env[62753]: DEBUG nova.compute.manager [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 690.256027] env[62753]: INFO nova.compute.manager [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: a408799d-29a9-442f-9769-8f8438af8ae0] Took 1.04 seconds to deallocate network for instance. [ 690.610185] env[62753]: DEBUG oslo_concurrency.lockutils [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.551s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 690.610705] env[62753]: DEBUG nova.compute.manager [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 690.613705] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 15.204s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 690.784181] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.794168] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "7563bf4e-39c5-4831-a415-cb311e237762" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 690.794661] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "7563bf4e-39c5-4831-a415-cb311e237762" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 691.117137] env[62753]: DEBUG nova.compute.utils [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 691.118605] env[62753]: DEBUG nova.compute.manager [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 691.122021] env[62753]: DEBUG nova.network.neutron [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 691.167605] env[62753]: DEBUG nova.policy [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '512275977cb04828a698cd3814808c61', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '3155e6eb2110482ab5dc76e87207dbba', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 691.308159] env[62753]: INFO nova.scheduler.client.report [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Deleted allocations for instance a408799d-29a9-442f-9769-8f8438af8ae0 [ 691.547327] env[62753]: DEBUG nova.network.neutron [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Successfully created port: e0e7e24b-752f-4598-b03e-3ac3e1ec3141 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 691.626427] env[62753]: DEBUG nova.compute.manager [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 691.659677] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance d81d5b85-fe31-43ec-91f6-7ff3251ff0c7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 691.659780] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance f10ff402-4308-4c75-af79-26e2942e0b92 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 691.660399] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 5fcbd37f-0568-43db-ae44-596eb4467391 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 691.820138] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d324b604-7cf6-4006-a84b-1bf5bf957f9e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "a408799d-29a9-442f-9769-8f8438af8ae0" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 91.796s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 692.165357] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance efaf8c7c-3f38-4a18-97e8-099da390c7d8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.323556] env[62753]: DEBUG nova.compute.manager [None req-67f953b7-d25d-4ea2-b4af-434ce4585a43 tempest-ServersListShow296Test-1505694806 tempest-ServersListShow296Test-1505694806-project-member] [instance: f0d83743-7a3b-4627-b69e-ac5c8e254a0e] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 692.421448] env[62753]: DEBUG nova.compute.manager [req-d39c9e95-0dd8-4e18-9658-ef62afb1d3a7 req-ac1c1296-a122-4ad2-acdb-8d93a11991ce service nova] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Received event network-changed-e0e7e24b-752f-4598-b03e-3ac3e1ec3141 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 692.421528] env[62753]: DEBUG nova.compute.manager [req-d39c9e95-0dd8-4e18-9658-ef62afb1d3a7 req-ac1c1296-a122-4ad2-acdb-8d93a11991ce service nova] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Refreshing instance network info cache due to event network-changed-e0e7e24b-752f-4598-b03e-3ac3e1ec3141. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 692.422434] env[62753]: DEBUG oslo_concurrency.lockutils [req-d39c9e95-0dd8-4e18-9658-ef62afb1d3a7 req-ac1c1296-a122-4ad2-acdb-8d93a11991ce service nova] Acquiring lock "refresh_cache-5fcbd37f-0568-43db-ae44-596eb4467391" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.422434] env[62753]: DEBUG oslo_concurrency.lockutils [req-d39c9e95-0dd8-4e18-9658-ef62afb1d3a7 req-ac1c1296-a122-4ad2-acdb-8d93a11991ce service nova] Acquired lock "refresh_cache-5fcbd37f-0568-43db-ae44-596eb4467391" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 692.422434] env[62753]: DEBUG nova.network.neutron [req-d39c9e95-0dd8-4e18-9658-ef62afb1d3a7 req-ac1c1296-a122-4ad2-acdb-8d93a11991ce service nova] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Refreshing network info cache for port e0e7e24b-752f-4598-b03e-3ac3e1ec3141 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 692.546926] env[62753]: ERROR nova.compute.manager [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port e0e7e24b-752f-4598-b03e-3ac3e1ec3141, please check neutron logs for more information. [ 692.546926] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 692.546926] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.546926] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 692.546926] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.546926] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 692.546926] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.546926] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 692.546926] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.546926] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 692.546926] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.546926] env[62753]: ERROR nova.compute.manager raise self.value [ 692.546926] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.546926] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 692.546926] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.546926] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 692.547341] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.547341] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 692.547341] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port e0e7e24b-752f-4598-b03e-3ac3e1ec3141, please check neutron logs for more information. [ 692.547341] env[62753]: ERROR nova.compute.manager [ 692.547341] env[62753]: Traceback (most recent call last): [ 692.547341] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 692.547341] env[62753]: listener.cb(fileno) [ 692.547341] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.547341] env[62753]: result = function(*args, **kwargs) [ 692.547341] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 692.547341] env[62753]: return func(*args, **kwargs) [ 692.547341] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.547341] env[62753]: raise e [ 692.547341] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.547341] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 692.547341] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.547341] env[62753]: created_port_ids = self._update_ports_for_instance( [ 692.547341] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.547341] env[62753]: with excutils.save_and_reraise_exception(): [ 692.547341] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.547341] env[62753]: self.force_reraise() [ 692.547341] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.547341] env[62753]: raise self.value [ 692.547341] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.547341] env[62753]: updated_port = self._update_port( [ 692.547341] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.547341] env[62753]: _ensure_no_port_binding_failure(port) [ 692.547341] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.547341] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 692.547980] env[62753]: nova.exception.PortBindingFailed: Binding failed for port e0e7e24b-752f-4598-b03e-3ac3e1ec3141, please check neutron logs for more information. [ 692.547980] env[62753]: Removing descriptor: 16 [ 692.643974] env[62753]: DEBUG nova.compute.manager [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 692.671135] env[62753]: DEBUG nova.virt.hardware [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 692.671393] env[62753]: DEBUG nova.virt.hardware [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 692.671551] env[62753]: DEBUG nova.virt.hardware [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 692.671732] env[62753]: DEBUG nova.virt.hardware [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 692.672026] env[62753]: DEBUG nova.virt.hardware [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 692.672114] env[62753]: DEBUG nova.virt.hardware [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 692.672328] env[62753]: DEBUG nova.virt.hardware [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 692.672488] env[62753]: DEBUG nova.virt.hardware [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 692.672656] env[62753]: DEBUG nova.virt.hardware [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 692.673240] env[62753]: DEBUG nova.virt.hardware [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 692.673240] env[62753]: DEBUG nova.virt.hardware [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 692.673755] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 3f46fa44-4410-4ca9-98d0-8efa4cf06d52 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 692.675550] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8152daec-4af9-42f2-b575-1a638dc90684 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.684504] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bd784bd-9754-402f-835d-fe2c33f967df {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 692.700370] env[62753]: ERROR nova.compute.manager [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port e0e7e24b-752f-4598-b03e-3ac3e1ec3141, please check neutron logs for more information. [ 692.700370] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Traceback (most recent call last): [ 692.700370] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 692.700370] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] yield resources [ 692.700370] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 692.700370] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] self.driver.spawn(context, instance, image_meta, [ 692.700370] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 692.700370] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] self._vmops.spawn(context, instance, image_meta, injected_files, [ 692.700370] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 692.700370] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] vm_ref = self.build_virtual_machine(instance, [ 692.700370] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 692.700748] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] vif_infos = vmwarevif.get_vif_info(self._session, [ 692.700748] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 692.700748] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] for vif in network_info: [ 692.700748] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 692.700748] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] return self._sync_wrapper(fn, *args, **kwargs) [ 692.700748] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 692.700748] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] self.wait() [ 692.700748] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 692.700748] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] self[:] = self._gt.wait() [ 692.700748] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 692.700748] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] return self._exit_event.wait() [ 692.700748] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 692.700748] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] current.throw(*self._exc) [ 692.701115] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 692.701115] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] result = function(*args, **kwargs) [ 692.701115] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 692.701115] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] return func(*args, **kwargs) [ 692.701115] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 692.701115] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] raise e [ 692.701115] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 692.701115] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] nwinfo = self.network_api.allocate_for_instance( [ 692.701115] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 692.701115] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] created_port_ids = self._update_ports_for_instance( [ 692.701115] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 692.701115] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] with excutils.save_and_reraise_exception(): [ 692.701115] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 692.701425] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] self.force_reraise() [ 692.701425] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 692.701425] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] raise self.value [ 692.701425] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 692.701425] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] updated_port = self._update_port( [ 692.701425] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 692.701425] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] _ensure_no_port_binding_failure(port) [ 692.701425] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 692.701425] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] raise exception.PortBindingFailed(port_id=port['id']) [ 692.701425] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] nova.exception.PortBindingFailed: Binding failed for port e0e7e24b-752f-4598-b03e-3ac3e1ec3141, please check neutron logs for more information. [ 692.701425] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] [ 692.701425] env[62753]: INFO nova.compute.manager [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Terminating instance [ 692.702844] env[62753]: DEBUG oslo_concurrency.lockutils [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Acquiring lock "refresh_cache-5fcbd37f-0568-43db-ae44-596eb4467391" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 692.829577] env[62753]: DEBUG nova.compute.manager [None req-67f953b7-d25d-4ea2-b4af-434ce4585a43 tempest-ServersListShow296Test-1505694806 tempest-ServersListShow296Test-1505694806-project-member] [instance: f0d83743-7a3b-4627-b69e-ac5c8e254a0e] Instance disappeared before build. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2439}} [ 692.886587] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Acquiring lock "07c4e02b-a8ca-41bb-82ec-38a81b9f1d07" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 692.886587] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Lock "07c4e02b-a8ca-41bb-82ec-38a81b9f1d07" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 692.943036] env[62753]: DEBUG nova.network.neutron [req-d39c9e95-0dd8-4e18-9658-ef62afb1d3a7 req-ac1c1296-a122-4ad2-acdb-8d93a11991ce service nova] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 693.075972] env[62753]: DEBUG nova.network.neutron [req-d39c9e95-0dd8-4e18-9658-ef62afb1d3a7 req-ac1c1296-a122-4ad2-acdb-8d93a11991ce service nova] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 693.181618] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 5ef5c03c-6c22-4db6-95ad-6c407c0bca78 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.342091] env[62753]: DEBUG oslo_concurrency.lockutils [None req-67f953b7-d25d-4ea2-b4af-434ce4585a43 tempest-ServersListShow296Test-1505694806 tempest-ServersListShow296Test-1505694806-project-member] Lock "f0d83743-7a3b-4627-b69e-ac5c8e254a0e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 62.429s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 693.583572] env[62753]: DEBUG oslo_concurrency.lockutils [req-d39c9e95-0dd8-4e18-9658-ef62afb1d3a7 req-ac1c1296-a122-4ad2-acdb-8d93a11991ce service nova] Releasing lock "refresh_cache-5fcbd37f-0568-43db-ae44-596eb4467391" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 693.583572] env[62753]: DEBUG oslo_concurrency.lockutils [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Acquired lock "refresh_cache-5fcbd37f-0568-43db-ae44-596eb4467391" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 693.583572] env[62753]: DEBUG nova.network.neutron [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 693.686139] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 50a480ed-f320-45d2-8ac9-a42d01614d11 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 693.846043] env[62753]: DEBUG nova.compute.manager [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 694.103176] env[62753]: DEBUG nova.network.neutron [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 694.194769] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 250f2711-6f32-4c4d-9876-35a3a1c30d12 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 694.196699] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "67f5d96c-b870-442a-a0f3-a70533c045ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.196924] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "67f5d96c-b870-442a-a0f3-a70533c045ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 694.233570] env[62753]: DEBUG nova.network.neutron [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 694.363256] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 694.472463] env[62753]: DEBUG nova.compute.manager [req-2952c710-e169-421c-b1ea-1a99e1d17bca req-777ef18e-87c7-4a91-a13b-eb084bfe64f0 service nova] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Received event network-vif-deleted-e0e7e24b-752f-4598-b03e-3ac3e1ec3141 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 694.700572] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance a413174f-2a97-4c41-b512-e8ab562a25f4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 694.736185] env[62753]: DEBUG oslo_concurrency.lockutils [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Releasing lock "refresh_cache-5fcbd37f-0568-43db-ae44-596eb4467391" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 694.736678] env[62753]: DEBUG nova.compute.manager [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 694.736943] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 694.737443] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-aab2aaad-c052-43fb-9021-9db3160af732 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.752025] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5db4f49e-4aee-48a0-81ce-b1362e1162f3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 694.773438] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5fcbd37f-0568-43db-ae44-596eb4467391 could not be found. [ 694.773639] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 694.774265] env[62753]: INFO nova.compute.manager [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Took 0.04 seconds to destroy the instance on the hypervisor. [ 694.774265] env[62753]: DEBUG oslo.service.loopingcall [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 694.774265] env[62753]: DEBUG nova.compute.manager [-] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 694.774546] env[62753]: DEBUG nova.network.neutron [-] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 694.793603] env[62753]: DEBUG nova.network.neutron [-] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 695.207638] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance c4ccc036-4963-4499-8bbb-dad5c0e4687e has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.296030] env[62753]: DEBUG nova.network.neutron [-] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 695.715132] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance a8ceac07-6631-4ac6-b62a-7f3eac788f2b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 695.799136] env[62753]: INFO nova.compute.manager [-] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Took 1.02 seconds to deallocate network for instance. [ 695.802690] env[62753]: DEBUG nova.compute.claims [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 695.802871] env[62753]: DEBUG oslo_concurrency.lockutils [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.217885] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance da9b45e9-618d-4e1e-9c49-3d702b2435e9 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 696.278190] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Acquiring lock "92e86d6a-e455-4984-9fdb-d1cdf856e729" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 696.278428] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Lock "92e86d6a-e455-4984-9fdb-d1cdf856e729" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 696.721142] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance e755857a-e0bc-40bf-bcbb-a416d243442b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.225655] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance eca070f0-dda0-4c51-b2d3-1251c517804a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 697.728954] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 698.232736] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 358d1c98-28ee-4f18-bd42-89a1e7687d03 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 698.736639] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 365fe802-f5e1-47d6-807e-a3b1027def02 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 699.238927] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 8664ed5c-ecb9-4795-8499-31198cfd0450 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 699.742347] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 700.246055] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance e260eaf7-98fe-461a-bc69-fc42dfcb79a8 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 700.748479] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 701.251649] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 701.754262] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance e0c4a1d2-6a85-4826-88fa-d21556d744ac has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 702.257786] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 41476adb-95e5-4617-b178-9c981d13c43d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 702.763108] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 3adb208d-cf70-4d93-b034-a3e87d029f90 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 703.266895] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance f63a8912-0abf-4c76-aaae-4085b3909688 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 703.770762] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 7563bf4e-39c5-4831-a415-cb311e237762 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 703.771039] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Total usable vcpus: 48, total allocated vcpus: 3 {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 703.772637] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1088MB phys_disk=200GB used_disk=3GB total_vcpus=48 used_vcpus=3 pci_stats=[] {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 704.084063] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e5e877f-4486-44a4-a9ee-3c442e82c09b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.091207] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0545a40-08a4-40e1-8de1-3be21dbdbede {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.119283] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8dd9234c-84ca-406a-92a7-5862841d6711 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.125802] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e31d1ce-5e26-4f57-9d50-389581f42fa9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 704.138272] env[62753]: DEBUG nova.compute.provider_tree [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 704.642073] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 705.147796] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 705.148132] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 14.534s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 705.148350] env[62753]: DEBUG oslo_concurrency.lockutils [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.469s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 705.149886] env[62753]: INFO nova.compute.claims [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 706.485210] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e305809c-ea71-4851-aaf2-f79292851295 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.492500] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c8b895d-3450-4157-a8b9-16e86446605c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.522850] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebfc9143-5da4-4737-80a1-e92e8de42bd8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.530366] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1edfbca1-7231-4c29-9989-79d31d2cb136 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 706.543403] env[62753]: DEBUG nova.compute.provider_tree [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 707.046842] env[62753]: DEBUG nova.scheduler.client.report [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 707.553046] env[62753]: DEBUG oslo_concurrency.lockutils [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.404s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 707.553328] env[62753]: DEBUG nova.compute.manager [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 707.555757] env[62753]: DEBUG oslo_concurrency.lockutils [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.969s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 707.557236] env[62753]: INFO nova.compute.claims [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 708.063187] env[62753]: DEBUG nova.compute.utils [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 708.064730] env[62753]: DEBUG nova.compute.manager [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 708.064923] env[62753]: DEBUG nova.network.neutron [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 708.107229] env[62753]: DEBUG nova.policy [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '426580c5caf34a14a2f1efd1f532f749', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0565c61333ee4a799f01e6cf3bde7ea5', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 708.347730] env[62753]: DEBUG nova.network.neutron [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Successfully created port: ff4576be-cfd3-4fd8-b156-f822a6bf1651 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 708.565532] env[62753]: DEBUG nova.compute.manager [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 708.942025] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6dc6647-edbd-42af-b974-501f56a1c972 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.946666] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d0ffb3c-7115-4c6a-9ad0-2b0ddb12a345 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.975695] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b73b58d-26f5-4022-b2d9-7fc5b8e768a5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.983115] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-004dfe7a-2ecc-43d3-b135-560f463fd03b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 708.996070] env[62753]: DEBUG nova.compute.provider_tree [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 709.209834] env[62753]: DEBUG nova.compute.manager [req-c9594cd4-3528-4bef-8774-c16824b96082 req-a0d55de0-c21a-410d-8c1e-45a2958630c8 service nova] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Received event network-changed-ff4576be-cfd3-4fd8-b156-f822a6bf1651 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 709.210051] env[62753]: DEBUG nova.compute.manager [req-c9594cd4-3528-4bef-8774-c16824b96082 req-a0d55de0-c21a-410d-8c1e-45a2958630c8 service nova] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Refreshing instance network info cache due to event network-changed-ff4576be-cfd3-4fd8-b156-f822a6bf1651. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 709.210271] env[62753]: DEBUG oslo_concurrency.lockutils [req-c9594cd4-3528-4bef-8774-c16824b96082 req-a0d55de0-c21a-410d-8c1e-45a2958630c8 service nova] Acquiring lock "refresh_cache-efaf8c7c-3f38-4a18-97e8-099da390c7d8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.210416] env[62753]: DEBUG oslo_concurrency.lockutils [req-c9594cd4-3528-4bef-8774-c16824b96082 req-a0d55de0-c21a-410d-8c1e-45a2958630c8 service nova] Acquired lock "refresh_cache-efaf8c7c-3f38-4a18-97e8-099da390c7d8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 709.210575] env[62753]: DEBUG nova.network.neutron [req-c9594cd4-3528-4bef-8774-c16824b96082 req-a0d55de0-c21a-410d-8c1e-45a2958630c8 service nova] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Refreshing network info cache for port ff4576be-cfd3-4fd8-b156-f822a6bf1651 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 709.310217] env[62753]: ERROR nova.compute.manager [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ff4576be-cfd3-4fd8-b156-f822a6bf1651, please check neutron logs for more information. [ 709.310217] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 709.310217] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.310217] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 709.310217] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.310217] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 709.310217] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.310217] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 709.310217] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.310217] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 709.310217] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.310217] env[62753]: ERROR nova.compute.manager raise self.value [ 709.310217] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.310217] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 709.310217] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.310217] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 709.310727] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.310727] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 709.310727] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ff4576be-cfd3-4fd8-b156-f822a6bf1651, please check neutron logs for more information. [ 709.310727] env[62753]: ERROR nova.compute.manager [ 709.310727] env[62753]: Traceback (most recent call last): [ 709.310727] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 709.310727] env[62753]: listener.cb(fileno) [ 709.310727] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.310727] env[62753]: result = function(*args, **kwargs) [ 709.310727] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 709.310727] env[62753]: return func(*args, **kwargs) [ 709.310727] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 709.310727] env[62753]: raise e [ 709.310727] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.310727] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 709.310727] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.310727] env[62753]: created_port_ids = self._update_ports_for_instance( [ 709.310727] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.310727] env[62753]: with excutils.save_and_reraise_exception(): [ 709.310727] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.310727] env[62753]: self.force_reraise() [ 709.310727] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.310727] env[62753]: raise self.value [ 709.310727] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.310727] env[62753]: updated_port = self._update_port( [ 709.310727] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.310727] env[62753]: _ensure_no_port_binding_failure(port) [ 709.310727] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.310727] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 709.311452] env[62753]: nova.exception.PortBindingFailed: Binding failed for port ff4576be-cfd3-4fd8-b156-f822a6bf1651, please check neutron logs for more information. [ 709.311452] env[62753]: Removing descriptor: 16 [ 709.499025] env[62753]: DEBUG nova.scheduler.client.report [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 709.574867] env[62753]: DEBUG nova.compute.manager [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 709.601544] env[62753]: DEBUG nova.virt.hardware [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 709.601912] env[62753]: DEBUG nova.virt.hardware [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 709.602047] env[62753]: DEBUG nova.virt.hardware [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 709.602288] env[62753]: DEBUG nova.virt.hardware [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 709.602472] env[62753]: DEBUG nova.virt.hardware [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 709.602656] env[62753]: DEBUG nova.virt.hardware [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 709.602902] env[62753]: DEBUG nova.virt.hardware [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 709.603116] env[62753]: DEBUG nova.virt.hardware [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 709.603332] env[62753]: DEBUG nova.virt.hardware [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 709.603588] env[62753]: DEBUG nova.virt.hardware [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 709.603828] env[62753]: DEBUG nova.virt.hardware [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 709.605214] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-370ec321-33b0-4980-b4df-a76478a58829 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.613048] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11043d8d-1219-4e24-94b2-ea960389f8d6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 709.627286] env[62753]: ERROR nova.compute.manager [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ff4576be-cfd3-4fd8-b156-f822a6bf1651, please check neutron logs for more information. [ 709.627286] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Traceback (most recent call last): [ 709.627286] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 709.627286] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] yield resources [ 709.627286] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 709.627286] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] self.driver.spawn(context, instance, image_meta, [ 709.627286] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 709.627286] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 709.627286] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 709.627286] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] vm_ref = self.build_virtual_machine(instance, [ 709.627286] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 709.627629] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] vif_infos = vmwarevif.get_vif_info(self._session, [ 709.627629] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 709.627629] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] for vif in network_info: [ 709.627629] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 709.627629] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] return self._sync_wrapper(fn, *args, **kwargs) [ 709.627629] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 709.627629] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] self.wait() [ 709.627629] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 709.627629] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] self[:] = self._gt.wait() [ 709.627629] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 709.627629] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] return self._exit_event.wait() [ 709.627629] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 709.627629] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] current.throw(*self._exc) [ 709.627974] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 709.627974] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] result = function(*args, **kwargs) [ 709.627974] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 709.627974] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] return func(*args, **kwargs) [ 709.627974] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 709.627974] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] raise e [ 709.627974] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 709.627974] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] nwinfo = self.network_api.allocate_for_instance( [ 709.627974] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 709.627974] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] created_port_ids = self._update_ports_for_instance( [ 709.627974] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 709.627974] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] with excutils.save_and_reraise_exception(): [ 709.627974] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 709.628308] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] self.force_reraise() [ 709.628308] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 709.628308] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] raise self.value [ 709.628308] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 709.628308] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] updated_port = self._update_port( [ 709.628308] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 709.628308] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] _ensure_no_port_binding_failure(port) [ 709.628308] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 709.628308] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] raise exception.PortBindingFailed(port_id=port['id']) [ 709.628308] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] nova.exception.PortBindingFailed: Binding failed for port ff4576be-cfd3-4fd8-b156-f822a6bf1651, please check neutron logs for more information. [ 709.628308] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] [ 709.628308] env[62753]: INFO nova.compute.manager [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Terminating instance [ 709.631296] env[62753]: DEBUG oslo_concurrency.lockutils [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Acquiring lock "refresh_cache-efaf8c7c-3f38-4a18-97e8-099da390c7d8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 709.730225] env[62753]: DEBUG nova.network.neutron [req-c9594cd4-3528-4bef-8774-c16824b96082 req-a0d55de0-c21a-410d-8c1e-45a2958630c8 service nova] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 709.808255] env[62753]: DEBUG nova.network.neutron [req-c9594cd4-3528-4bef-8774-c16824b96082 req-a0d55de0-c21a-410d-8c1e-45a2958630c8 service nova] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.004713] env[62753]: DEBUG oslo_concurrency.lockutils [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.448s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 710.004713] env[62753]: DEBUG nova.compute.manager [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 710.007274] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.092s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 710.310933] env[62753]: DEBUG oslo_concurrency.lockutils [req-c9594cd4-3528-4bef-8774-c16824b96082 req-a0d55de0-c21a-410d-8c1e-45a2958630c8 service nova] Releasing lock "refresh_cache-efaf8c7c-3f38-4a18-97e8-099da390c7d8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 710.311080] env[62753]: DEBUG oslo_concurrency.lockutils [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Acquired lock "refresh_cache-efaf8c7c-3f38-4a18-97e8-099da390c7d8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 710.311247] env[62753]: DEBUG nova.network.neutron [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 710.511975] env[62753]: DEBUG nova.compute.utils [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 710.519462] env[62753]: DEBUG nova.compute.manager [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 710.519462] env[62753]: DEBUG nova.network.neutron [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 710.558750] env[62753]: DEBUG nova.policy [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9931b40cfd7846038805c6a4caedaac0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b8e81660b30408c998e412f5fa81469', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 710.843402] env[62753]: DEBUG nova.network.neutron [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Successfully created port: ff6cde93-e89f-4cec-9dc3-50d57ab15e2f {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 710.903459] env[62753]: DEBUG nova.network.neutron [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 710.969517] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f32553d-1702-40c4-ae9c-0181ecb991f8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 710.983127] env[62753]: DEBUG nova.network.neutron [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 710.985515] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63d3de98-a47f-4452-bda7-4ac196adf3e8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.019789] env[62753]: DEBUG nova.compute.manager [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 711.024202] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-194d60ee-35ae-4a2f-9366-3e9ead525441 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.032107] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03603ddd-7d65-4745-bb37-bb2d8a56a30c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.047920] env[62753]: DEBUG nova.compute.provider_tree [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 711.263172] env[62753]: DEBUG nova.compute.manager [req-2a0061de-f8a0-4ede-8455-8be7177a9817 req-9ec400c9-6418-4d0a-b129-23597589bfc7 service nova] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Received event network-vif-deleted-ff4576be-cfd3-4fd8-b156-f822a6bf1651 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 711.490095] env[62753]: DEBUG oslo_concurrency.lockutils [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Releasing lock "refresh_cache-efaf8c7c-3f38-4a18-97e8-099da390c7d8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 711.490493] env[62753]: DEBUG nova.compute.manager [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 711.490697] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 711.491033] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-115c72e2-80b5-4cad-a248-e722550e1668 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.498793] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c9ad9b-b4d4-4436-9317-b64ec048af6c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 711.519963] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance efaf8c7c-3f38-4a18-97e8-099da390c7d8 could not be found. [ 711.520178] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 711.520363] env[62753]: INFO nova.compute.manager [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Took 0.03 seconds to destroy the instance on the hypervisor. [ 711.520593] env[62753]: DEBUG oslo.service.loopingcall [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 711.520791] env[62753]: DEBUG nova.compute.manager [-] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 711.520885] env[62753]: DEBUG nova.network.neutron [-] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 711.543145] env[62753]: DEBUG nova.network.neutron [-] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 711.551378] env[62753]: DEBUG nova.scheduler.client.report [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 711.693661] env[62753]: ERROR nova.compute.manager [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port ff6cde93-e89f-4cec-9dc3-50d57ab15e2f, please check neutron logs for more information. [ 711.693661] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 711.693661] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.693661] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 711.693661] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.693661] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 711.693661] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.693661] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 711.693661] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.693661] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 711.693661] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.693661] env[62753]: ERROR nova.compute.manager raise self.value [ 711.693661] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.693661] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 711.693661] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.693661] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 711.694126] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.694126] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 711.694126] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port ff6cde93-e89f-4cec-9dc3-50d57ab15e2f, please check neutron logs for more information. [ 711.694126] env[62753]: ERROR nova.compute.manager [ 711.694126] env[62753]: Traceback (most recent call last): [ 711.694126] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 711.694126] env[62753]: listener.cb(fileno) [ 711.694126] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 711.694126] env[62753]: result = function(*args, **kwargs) [ 711.694126] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 711.694126] env[62753]: return func(*args, **kwargs) [ 711.694126] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 711.694126] env[62753]: raise e [ 711.694126] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 711.694126] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 711.694126] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 711.694126] env[62753]: created_port_ids = self._update_ports_for_instance( [ 711.694126] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 711.694126] env[62753]: with excutils.save_and_reraise_exception(): [ 711.694126] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 711.694126] env[62753]: self.force_reraise() [ 711.694126] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 711.694126] env[62753]: raise self.value [ 711.694126] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 711.694126] env[62753]: updated_port = self._update_port( [ 711.694126] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 711.694126] env[62753]: _ensure_no_port_binding_failure(port) [ 711.694126] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 711.694126] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 711.695048] env[62753]: nova.exception.PortBindingFailed: Binding failed for port ff6cde93-e89f-4cec-9dc3-50d57ab15e2f, please check neutron logs for more information. [ 711.695048] env[62753]: Removing descriptor: 16 [ 712.032015] env[62753]: DEBUG nova.compute.manager [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 712.044877] env[62753]: DEBUG nova.network.neutron [-] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.057066] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.050s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 712.057988] env[62753]: ERROR nova.compute.manager [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 655873cb-9429-4d6c-8a0a-27b6cfabd148, please check neutron logs for more information. [ 712.057988] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Traceback (most recent call last): [ 712.057988] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.057988] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] self.driver.spawn(context, instance, image_meta, [ 712.057988] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 712.057988] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.057988] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.057988] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] vm_ref = self.build_virtual_machine(instance, [ 712.057988] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.057988] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.057988] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.058396] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] for vif in network_info: [ 712.058396] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.058396] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] return self._sync_wrapper(fn, *args, **kwargs) [ 712.058396] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.058396] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] self.wait() [ 712.058396] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.058396] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] self[:] = self._gt.wait() [ 712.058396] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.058396] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] return self._exit_event.wait() [ 712.058396] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 712.058396] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] result = hub.switch() [ 712.058396] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 712.058396] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] return self.greenlet.switch() [ 712.058766] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.058766] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] result = function(*args, **kwargs) [ 712.058766] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 712.058766] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] return func(*args, **kwargs) [ 712.058766] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.058766] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] raise e [ 712.058766] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.058766] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] nwinfo = self.network_api.allocate_for_instance( [ 712.058766] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.058766] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] created_port_ids = self._update_ports_for_instance( [ 712.058766] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.058766] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] with excutils.save_and_reraise_exception(): [ 712.058766] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.059254] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] self.force_reraise() [ 712.059254] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.059254] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] raise self.value [ 712.059254] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.059254] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] updated_port = self._update_port( [ 712.059254] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.059254] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] _ensure_no_port_binding_failure(port) [ 712.059254] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.059254] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] raise exception.PortBindingFailed(port_id=port['id']) [ 712.059254] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] nova.exception.PortBindingFailed: Binding failed for port 655873cb-9429-4d6c-8a0a-27b6cfabd148, please check neutron logs for more information. [ 712.059254] env[62753]: ERROR nova.compute.manager [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] [ 712.059596] env[62753]: DEBUG nova.compute.utils [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Binding failed for port 655873cb-9429-4d6c-8a0a-27b6cfabd148, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 712.062208] env[62753]: DEBUG nova.virt.hardware [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 712.062425] env[62753]: DEBUG nova.virt.hardware [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 712.062578] env[62753]: DEBUG nova.virt.hardware [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 712.062756] env[62753]: DEBUG nova.virt.hardware [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 712.062902] env[62753]: DEBUG nova.virt.hardware [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 712.063061] env[62753]: DEBUG nova.virt.hardware [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 712.063287] env[62753]: DEBUG nova.virt.hardware [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 712.063460] env[62753]: DEBUG nova.virt.hardware [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 712.063631] env[62753]: DEBUG nova.virt.hardware [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 712.063794] env[62753]: DEBUG nova.virt.hardware [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 712.063963] env[62753]: DEBUG nova.virt.hardware [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 712.064291] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.923s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 712.066055] env[62753]: INFO nova.compute.claims [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 712.068795] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb011252-ba23-4959-9eba-dbf2b7f9fd63 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.071967] env[62753]: DEBUG nova.compute.manager [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Build of instance d81d5b85-fe31-43ec-91f6-7ff3251ff0c7 was re-scheduled: Binding failed for port 655873cb-9429-4d6c-8a0a-27b6cfabd148, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 712.072439] env[62753]: DEBUG nova.compute.manager [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 712.072670] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Acquiring lock "refresh_cache-d81d5b85-fe31-43ec-91f6-7ff3251ff0c7" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.072801] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Acquired lock "refresh_cache-d81d5b85-fe31-43ec-91f6-7ff3251ff0c7" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.072956] env[62753]: DEBUG nova.network.neutron [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 712.081342] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19168091-deb6-49ee-b5cd-f89dc12c04a3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 712.097041] env[62753]: ERROR nova.compute.manager [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port ff6cde93-e89f-4cec-9dc3-50d57ab15e2f, please check neutron logs for more information. [ 712.097041] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Traceback (most recent call last): [ 712.097041] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 712.097041] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] yield resources [ 712.097041] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 712.097041] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] self.driver.spawn(context, instance, image_meta, [ 712.097041] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 712.097041] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] self._vmops.spawn(context, instance, image_meta, injected_files, [ 712.097041] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 712.097041] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] vm_ref = self.build_virtual_machine(instance, [ 712.097041] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 712.097420] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] vif_infos = vmwarevif.get_vif_info(self._session, [ 712.097420] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 712.097420] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] for vif in network_info: [ 712.097420] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 712.097420] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] return self._sync_wrapper(fn, *args, **kwargs) [ 712.097420] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 712.097420] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] self.wait() [ 712.097420] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 712.097420] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] self[:] = self._gt.wait() [ 712.097420] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 712.097420] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] return self._exit_event.wait() [ 712.097420] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 712.097420] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] current.throw(*self._exc) [ 712.097716] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 712.097716] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] result = function(*args, **kwargs) [ 712.097716] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 712.097716] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] return func(*args, **kwargs) [ 712.097716] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 712.097716] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] raise e [ 712.097716] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 712.097716] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] nwinfo = self.network_api.allocate_for_instance( [ 712.097716] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 712.097716] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] created_port_ids = self._update_ports_for_instance( [ 712.097716] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 712.097716] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] with excutils.save_and_reraise_exception(): [ 712.097716] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 712.098055] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] self.force_reraise() [ 712.098055] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 712.098055] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] raise self.value [ 712.098055] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 712.098055] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] updated_port = self._update_port( [ 712.098055] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 712.098055] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] _ensure_no_port_binding_failure(port) [ 712.098055] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 712.098055] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] raise exception.PortBindingFailed(port_id=port['id']) [ 712.098055] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] nova.exception.PortBindingFailed: Binding failed for port ff6cde93-e89f-4cec-9dc3-50d57ab15e2f, please check neutron logs for more information. [ 712.098055] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] [ 712.098055] env[62753]: INFO nova.compute.manager [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Terminating instance [ 712.100158] env[62753]: DEBUG oslo_concurrency.lockutils [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "refresh_cache-3f46fa44-4410-4ca9-98d0-8efa4cf06d52" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 712.100321] env[62753]: DEBUG oslo_concurrency.lockutils [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "refresh_cache-3f46fa44-4410-4ca9-98d0-8efa4cf06d52" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 712.100486] env[62753]: DEBUG nova.network.neutron [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 712.548073] env[62753]: INFO nova.compute.manager [-] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Took 1.03 seconds to deallocate network for instance. [ 712.549798] env[62753]: DEBUG nova.compute.claims [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 712.549978] env[62753]: DEBUG oslo_concurrency.lockutils [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 712.592743] env[62753]: DEBUG nova.network.neutron [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.618162] env[62753]: DEBUG nova.network.neutron [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 712.664398] env[62753]: DEBUG nova.network.neutron [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 712.718999] env[62753]: DEBUG nova.network.neutron [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.166872] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Releasing lock "refresh_cache-d81d5b85-fe31-43ec-91f6-7ff3251ff0c7" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.167154] env[62753]: DEBUG nova.compute.manager [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 713.167297] env[62753]: DEBUG nova.compute.manager [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.167463] env[62753]: DEBUG nova.network.neutron [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 713.183453] env[62753]: DEBUG nova.network.neutron [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.221228] env[62753]: DEBUG oslo_concurrency.lockutils [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "refresh_cache-3f46fa44-4410-4ca9-98d0-8efa4cf06d52" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 713.221656] env[62753]: DEBUG nova.compute.manager [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 713.221843] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 713.222146] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a524aeea-216e-455a-b8fe-e7a08e46872a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.232568] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8fd910-59ec-4ca6-ab69-ea5ea87f311f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.255919] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3f46fa44-4410-4ca9-98d0-8efa4cf06d52 could not be found. [ 713.256169] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 713.256361] env[62753]: INFO nova.compute.manager [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Took 0.03 seconds to destroy the instance on the hypervisor. [ 713.256600] env[62753]: DEBUG oslo.service.loopingcall [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 713.259115] env[62753]: DEBUG nova.compute.manager [-] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 713.259194] env[62753]: DEBUG nova.network.neutron [-] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 713.274153] env[62753]: DEBUG nova.network.neutron [-] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 713.463751] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c1cbac-3d0a-436e-b4ad-163ac5efc682 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.471335] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62e29deb-4bd0-47fa-8c70-57000026962b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.501680] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf977f15-436d-4fad-9038-7cf187a37a28 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.509093] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ec3c6e-7e81-4532-8877-e6d62a8638d4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 713.523831] env[62753]: DEBUG nova.compute.provider_tree [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 713.529922] env[62753]: DEBUG nova.compute.manager [req-f7a95ffa-1dba-45bd-9982-61ad3b5d9101 req-b06129c3-5b24-4890-b9e9-1d3a19c7b206 service nova] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Received event network-changed-ff6cde93-e89f-4cec-9dc3-50d57ab15e2f {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 713.530103] env[62753]: DEBUG nova.compute.manager [req-f7a95ffa-1dba-45bd-9982-61ad3b5d9101 req-b06129c3-5b24-4890-b9e9-1d3a19c7b206 service nova] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Refreshing instance network info cache due to event network-changed-ff6cde93-e89f-4cec-9dc3-50d57ab15e2f. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 713.530307] env[62753]: DEBUG oslo_concurrency.lockutils [req-f7a95ffa-1dba-45bd-9982-61ad3b5d9101 req-b06129c3-5b24-4890-b9e9-1d3a19c7b206 service nova] Acquiring lock "refresh_cache-3f46fa44-4410-4ca9-98d0-8efa4cf06d52" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 713.530439] env[62753]: DEBUG oslo_concurrency.lockutils [req-f7a95ffa-1dba-45bd-9982-61ad3b5d9101 req-b06129c3-5b24-4890-b9e9-1d3a19c7b206 service nova] Acquired lock "refresh_cache-3f46fa44-4410-4ca9-98d0-8efa4cf06d52" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 713.530588] env[62753]: DEBUG nova.network.neutron [req-f7a95ffa-1dba-45bd-9982-61ad3b5d9101 req-b06129c3-5b24-4890-b9e9-1d3a19c7b206 service nova] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Refreshing network info cache for port ff6cde93-e89f-4cec-9dc3-50d57ab15e2f {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 713.686299] env[62753]: DEBUG nova.network.neutron [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 713.776987] env[62753]: DEBUG nova.network.neutron [-] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.026985] env[62753]: DEBUG nova.scheduler.client.report [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 714.066090] env[62753]: DEBUG nova.network.neutron [req-f7a95ffa-1dba-45bd-9982-61ad3b5d9101 req-b06129c3-5b24-4890-b9e9-1d3a19c7b206 service nova] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 714.131136] env[62753]: DEBUG nova.network.neutron [req-f7a95ffa-1dba-45bd-9982-61ad3b5d9101 req-b06129c3-5b24-4890-b9e9-1d3a19c7b206 service nova] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 714.190065] env[62753]: INFO nova.compute.manager [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] [instance: d81d5b85-fe31-43ec-91f6-7ff3251ff0c7] Took 1.02 seconds to deallocate network for instance. [ 714.281049] env[62753]: INFO nova.compute.manager [-] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Took 1.02 seconds to deallocate network for instance. [ 714.282420] env[62753]: DEBUG nova.compute.claims [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 714.282600] env[62753]: DEBUG oslo_concurrency.lockutils [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 714.532752] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.468s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 714.533500] env[62753]: DEBUG nova.compute.manager [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 714.537131] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.415s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 714.538695] env[62753]: INFO nova.compute.claims [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 714.633650] env[62753]: DEBUG oslo_concurrency.lockutils [req-f7a95ffa-1dba-45bd-9982-61ad3b5d9101 req-b06129c3-5b24-4890-b9e9-1d3a19c7b206 service nova] Releasing lock "refresh_cache-3f46fa44-4410-4ca9-98d0-8efa4cf06d52" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 714.633923] env[62753]: DEBUG nova.compute.manager [req-f7a95ffa-1dba-45bd-9982-61ad3b5d9101 req-b06129c3-5b24-4890-b9e9-1d3a19c7b206 service nova] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Received event network-vif-deleted-ff6cde93-e89f-4cec-9dc3-50d57ab15e2f {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 715.043872] env[62753]: DEBUG nova.compute.utils [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 715.046746] env[62753]: DEBUG nova.compute.manager [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 715.046839] env[62753]: DEBUG nova.network.neutron [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 715.085341] env[62753]: DEBUG nova.policy [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '7ac0d7c1c32f46cbad1759b0a14e71bf', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '019da95350994831b5d3934c7d453689', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 715.217021] env[62753]: INFO nova.scheduler.client.report [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Deleted allocations for instance d81d5b85-fe31-43ec-91f6-7ff3251ff0c7 [ 715.443032] env[62753]: DEBUG nova.network.neutron [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Successfully created port: 78ea3c2f-4f6e-4bca-8b2b-d4a3ddd039ca {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 715.547342] env[62753]: DEBUG nova.compute.manager [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 715.725350] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a6fa1f4d-2c8a-4973-a214-d88346205462 tempest-InstanceActionsTestJSON-1619272812 tempest-InstanceActionsTestJSON-1619272812-project-member] Lock "d81d5b85-fe31-43ec-91f6-7ff3251ff0c7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.340s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 715.991083] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ad63f33-f1dc-4c0a-9414-6b548ef0da75 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 715.998926] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11adac0b-1425-4382-8e34-b608749eb625 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.028461] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2f1fd32-b258-4072-ac32-fbcc6ca8c021 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.035612] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa66d0c9-ea2b-4129-9e0e-51ad183557ba {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.050036] env[62753]: DEBUG nova.compute.provider_tree [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 716.231104] env[62753]: DEBUG nova.compute.manager [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 716.503208] env[62753]: DEBUG nova.compute.manager [req-2f432748-621f-48e5-ad8b-7a9b76e015b2 req-e452a7b1-8390-4577-890f-a8d37829fc29 service nova] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Received event network-changed-78ea3c2f-4f6e-4bca-8b2b-d4a3ddd039ca {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 716.503345] env[62753]: DEBUG nova.compute.manager [req-2f432748-621f-48e5-ad8b-7a9b76e015b2 req-e452a7b1-8390-4577-890f-a8d37829fc29 service nova] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Refreshing instance network info cache due to event network-changed-78ea3c2f-4f6e-4bca-8b2b-d4a3ddd039ca. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 716.503574] env[62753]: DEBUG oslo_concurrency.lockutils [req-2f432748-621f-48e5-ad8b-7a9b76e015b2 req-e452a7b1-8390-4577-890f-a8d37829fc29 service nova] Acquiring lock "refresh_cache-5ef5c03c-6c22-4db6-95ad-6c407c0bca78" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.503801] env[62753]: DEBUG oslo_concurrency.lockutils [req-2f432748-621f-48e5-ad8b-7a9b76e015b2 req-e452a7b1-8390-4577-890f-a8d37829fc29 service nova] Acquired lock "refresh_cache-5ef5c03c-6c22-4db6-95ad-6c407c0bca78" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 716.504943] env[62753]: DEBUG nova.network.neutron [req-2f432748-621f-48e5-ad8b-7a9b76e015b2 req-e452a7b1-8390-4577-890f-a8d37829fc29 service nova] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Refreshing network info cache for port 78ea3c2f-4f6e-4bca-8b2b-d4a3ddd039ca {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 716.553256] env[62753]: DEBUG nova.scheduler.client.report [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 716.561190] env[62753]: DEBUG nova.compute.manager [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 716.597748] env[62753]: DEBUG nova.virt.hardware [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 716.598088] env[62753]: DEBUG nova.virt.hardware [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 716.598304] env[62753]: DEBUG nova.virt.hardware [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 716.598571] env[62753]: DEBUG nova.virt.hardware [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 716.598722] env[62753]: DEBUG nova.virt.hardware [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 716.598913] env[62753]: DEBUG nova.virt.hardware [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 716.599184] env[62753]: DEBUG nova.virt.hardware [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 716.599395] env[62753]: DEBUG nova.virt.hardware [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 716.599609] env[62753]: DEBUG nova.virt.hardware [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 716.599815] env[62753]: DEBUG nova.virt.hardware [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 716.600041] env[62753]: DEBUG nova.virt.hardware [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 716.601247] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7af2758-098e-44a9-a69a-941a8fb856d2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.610606] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb5e3679-be27-4530-8f26-9fccfced80a6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 716.642360] env[62753]: ERROR nova.compute.manager [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 78ea3c2f-4f6e-4bca-8b2b-d4a3ddd039ca, please check neutron logs for more information. [ 716.642360] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 716.642360] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.642360] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 716.642360] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.642360] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 716.642360] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.642360] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 716.642360] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.642360] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 716.642360] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.642360] env[62753]: ERROR nova.compute.manager raise self.value [ 716.642360] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.642360] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 716.642360] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.642360] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 716.642884] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.642884] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 716.642884] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 78ea3c2f-4f6e-4bca-8b2b-d4a3ddd039ca, please check neutron logs for more information. [ 716.642884] env[62753]: ERROR nova.compute.manager [ 716.642884] env[62753]: Traceback (most recent call last): [ 716.642884] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 716.642884] env[62753]: listener.cb(fileno) [ 716.642884] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.642884] env[62753]: result = function(*args, **kwargs) [ 716.642884] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 716.642884] env[62753]: return func(*args, **kwargs) [ 716.642884] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.642884] env[62753]: raise e [ 716.642884] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.642884] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 716.642884] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.642884] env[62753]: created_port_ids = self._update_ports_for_instance( [ 716.642884] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.642884] env[62753]: with excutils.save_and_reraise_exception(): [ 716.642884] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.642884] env[62753]: self.force_reraise() [ 716.642884] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.642884] env[62753]: raise self.value [ 716.642884] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.642884] env[62753]: updated_port = self._update_port( [ 716.642884] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.642884] env[62753]: _ensure_no_port_binding_failure(port) [ 716.642884] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.642884] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 716.644375] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 78ea3c2f-4f6e-4bca-8b2b-d4a3ddd039ca, please check neutron logs for more information. [ 716.644375] env[62753]: Removing descriptor: 14 [ 716.644375] env[62753]: ERROR nova.compute.manager [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 78ea3c2f-4f6e-4bca-8b2b-d4a3ddd039ca, please check neutron logs for more information. [ 716.644375] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Traceback (most recent call last): [ 716.644375] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 716.644375] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] yield resources [ 716.644375] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 716.644375] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] self.driver.spawn(context, instance, image_meta, [ 716.644375] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 716.644375] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] self._vmops.spawn(context, instance, image_meta, injected_files, [ 716.644375] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 716.644375] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] vm_ref = self.build_virtual_machine(instance, [ 716.644705] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 716.644705] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] vif_infos = vmwarevif.get_vif_info(self._session, [ 716.644705] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 716.644705] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] for vif in network_info: [ 716.644705] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 716.644705] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] return self._sync_wrapper(fn, *args, **kwargs) [ 716.644705] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 716.644705] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] self.wait() [ 716.644705] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 716.644705] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] self[:] = self._gt.wait() [ 716.644705] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 716.644705] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] return self._exit_event.wait() [ 716.644705] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 716.645028] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] result = hub.switch() [ 716.645028] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 716.645028] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] return self.greenlet.switch() [ 716.645028] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 716.645028] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] result = function(*args, **kwargs) [ 716.645028] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 716.645028] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] return func(*args, **kwargs) [ 716.645028] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 716.645028] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] raise e [ 716.645028] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 716.645028] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] nwinfo = self.network_api.allocate_for_instance( [ 716.645028] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 716.645028] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] created_port_ids = self._update_ports_for_instance( [ 716.645379] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 716.645379] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] with excutils.save_and_reraise_exception(): [ 716.645379] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 716.645379] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] self.force_reraise() [ 716.645379] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 716.645379] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] raise self.value [ 716.645379] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 716.645379] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] updated_port = self._update_port( [ 716.645379] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 716.645379] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] _ensure_no_port_binding_failure(port) [ 716.645379] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 716.645379] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] raise exception.PortBindingFailed(port_id=port['id']) [ 716.645766] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] nova.exception.PortBindingFailed: Binding failed for port 78ea3c2f-4f6e-4bca-8b2b-d4a3ddd039ca, please check neutron logs for more information. [ 716.645766] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] [ 716.645766] env[62753]: INFO nova.compute.manager [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Terminating instance [ 716.645766] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Acquiring lock "refresh_cache-5ef5c03c-6c22-4db6-95ad-6c407c0bca78" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 716.758242] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 717.025129] env[62753]: DEBUG nova.network.neutron [req-2f432748-621f-48e5-ad8b-7a9b76e015b2 req-e452a7b1-8390-4577-890f-a8d37829fc29 service nova] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 717.059341] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.522s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 717.059886] env[62753]: DEBUG nova.compute.manager [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 717.062831] env[62753]: DEBUG oslo_concurrency.lockutils [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 29.382s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 717.161571] env[62753]: DEBUG nova.network.neutron [req-2f432748-621f-48e5-ad8b-7a9b76e015b2 req-e452a7b1-8390-4577-890f-a8d37829fc29 service nova] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 717.568878] env[62753]: DEBUG nova.compute.utils [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 717.573823] env[62753]: DEBUG nova.compute.manager [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 717.573823] env[62753]: DEBUG nova.network.neutron [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 717.639932] env[62753]: DEBUG nova.policy [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6ff3dbb3bdbc4d7db041de0341dcdff0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f5537c92669a4fc39fa27e67f0fb8816', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 717.663906] env[62753]: DEBUG oslo_concurrency.lockutils [req-2f432748-621f-48e5-ad8b-7a9b76e015b2 req-e452a7b1-8390-4577-890f-a8d37829fc29 service nova] Releasing lock "refresh_cache-5ef5c03c-6c22-4db6-95ad-6c407c0bca78" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 717.664323] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Acquired lock "refresh_cache-5ef5c03c-6c22-4db6-95ad-6c407c0bca78" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 717.664507] env[62753]: DEBUG nova.network.neutron [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 717.961804] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80ebc0d6-7fdd-409e-ac0f-f8adb70f500a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 717.969531] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb3ede8a-fe22-4e92-8fc1-5ccff11f642c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.002343] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c6df8f0-112d-4201-acf0-c12869d3dd85 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.011469] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fa9234b-18f2-43f1-82dd-cf962d307854 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.028546] env[62753]: DEBUG nova.compute.provider_tree [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 718.073940] env[62753]: DEBUG nova.compute.manager [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 718.188278] env[62753]: DEBUG nova.network.neutron [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 718.278892] env[62753]: DEBUG nova.network.neutron [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Successfully created port: c2f33405-b9de-41af-9eed-37383de235f0 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 718.281527] env[62753]: DEBUG nova.network.neutron [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 718.533610] env[62753]: DEBUG nova.scheduler.client.report [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 718.540384] env[62753]: DEBUG nova.compute.manager [req-5c226619-c96f-4f70-a580-bbf10c3d466c req-c19a4016-a092-4984-88cf-2e5b27eea2e2 service nova] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Received event network-vif-deleted-78ea3c2f-4f6e-4bca-8b2b-d4a3ddd039ca {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 718.784145] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Releasing lock "refresh_cache-5ef5c03c-6c22-4db6-95ad-6c407c0bca78" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 718.784586] env[62753]: DEBUG nova.compute.manager [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 718.784785] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 718.785097] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-75434cf3-8c8d-417b-bfe0-e31540fcfce4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.794010] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f0e48ab-b721-4665-8d5e-453f688c0f4c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 718.816301] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 5ef5c03c-6c22-4db6-95ad-6c407c0bca78 could not be found. [ 718.816301] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 718.816301] env[62753]: INFO nova.compute.manager [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Took 0.03 seconds to destroy the instance on the hypervisor. [ 718.816301] env[62753]: DEBUG oslo.service.loopingcall [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 718.816301] env[62753]: DEBUG nova.compute.manager [-] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 718.816301] env[62753]: DEBUG nova.network.neutron [-] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 718.836296] env[62753]: DEBUG nova.network.neutron [-] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.041837] env[62753]: DEBUG oslo_concurrency.lockutils [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.979s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 719.042499] env[62753]: ERROR nova.compute.manager [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port d93e8619-1348-4566-9c18-ebbebb6b1f46, please check neutron logs for more information. [ 719.042499] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Traceback (most recent call last): [ 719.042499] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 719.042499] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] self.driver.spawn(context, instance, image_meta, [ 719.042499] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 719.042499] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] self._vmops.spawn(context, instance, image_meta, injected_files, [ 719.042499] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 719.042499] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] vm_ref = self.build_virtual_machine(instance, [ 719.042499] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 719.042499] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] vif_infos = vmwarevif.get_vif_info(self._session, [ 719.042499] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 719.042897] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] for vif in network_info: [ 719.042897] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 719.042897] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] return self._sync_wrapper(fn, *args, **kwargs) [ 719.042897] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 719.042897] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] self.wait() [ 719.042897] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 719.042897] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] self[:] = self._gt.wait() [ 719.042897] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 719.042897] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] return self._exit_event.wait() [ 719.042897] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 719.042897] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] result = hub.switch() [ 719.042897] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 719.042897] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] return self.greenlet.switch() [ 719.043221] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.043221] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] result = function(*args, **kwargs) [ 719.043221] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 719.043221] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] return func(*args, **kwargs) [ 719.043221] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 719.043221] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] raise e [ 719.043221] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.043221] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] nwinfo = self.network_api.allocate_for_instance( [ 719.043221] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 719.043221] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] created_port_ids = self._update_ports_for_instance( [ 719.043221] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 719.043221] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] with excutils.save_and_reraise_exception(): [ 719.043221] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.043575] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] self.force_reraise() [ 719.043575] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.043575] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] raise self.value [ 719.043575] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 719.043575] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] updated_port = self._update_port( [ 719.043575] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.043575] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] _ensure_no_port_binding_failure(port) [ 719.043575] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.043575] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] raise exception.PortBindingFailed(port_id=port['id']) [ 719.043575] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] nova.exception.PortBindingFailed: Binding failed for port d93e8619-1348-4566-9c18-ebbebb6b1f46, please check neutron logs for more information. [ 719.043575] env[62753]: ERROR nova.compute.manager [instance: f10ff402-4308-4c75-af79-26e2942e0b92] [ 719.043840] env[62753]: DEBUG nova.compute.utils [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Binding failed for port d93e8619-1348-4566-9c18-ebbebb6b1f46, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 719.044473] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.287s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 719.045905] env[62753]: INFO nova.compute.claims [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 719.049357] env[62753]: DEBUG nova.compute.manager [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Build of instance f10ff402-4308-4c75-af79-26e2942e0b92 was re-scheduled: Binding failed for port d93e8619-1348-4566-9c18-ebbebb6b1f46, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 719.049357] env[62753]: DEBUG nova.compute.manager [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 719.049357] env[62753]: DEBUG oslo_concurrency.lockutils [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Acquiring lock "refresh_cache-f10ff402-4308-4c75-af79-26e2942e0b92" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.049357] env[62753]: DEBUG oslo_concurrency.lockutils [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Acquired lock "refresh_cache-f10ff402-4308-4c75-af79-26e2942e0b92" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.049568] env[62753]: DEBUG nova.network.neutron [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.084020] env[62753]: DEBUG nova.compute.manager [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 719.111027] env[62753]: DEBUG nova.virt.hardware [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 719.111027] env[62753]: DEBUG nova.virt.hardware [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 719.111027] env[62753]: DEBUG nova.virt.hardware [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 719.111322] env[62753]: DEBUG nova.virt.hardware [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 719.111467] env[62753]: DEBUG nova.virt.hardware [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 719.111665] env[62753]: DEBUG nova.virt.hardware [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 719.111952] env[62753]: DEBUG nova.virt.hardware [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 719.112167] env[62753]: DEBUG nova.virt.hardware [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 719.112407] env[62753]: DEBUG nova.virt.hardware [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 719.112631] env[62753]: DEBUG nova.virt.hardware [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 719.112845] env[62753]: DEBUG nova.virt.hardware [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 719.113758] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bd52f0b-acb6-47b1-b8c6-09da4db622d3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.122276] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8187acc-2bda-4c70-8377-86d4b3d9e293 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 719.339229] env[62753]: DEBUG nova.network.neutron [-] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.471099] env[62753]: ERROR nova.compute.manager [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c2f33405-b9de-41af-9eed-37383de235f0, please check neutron logs for more information. [ 719.471099] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 719.471099] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.471099] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 719.471099] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 719.471099] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 719.471099] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 719.471099] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 719.471099] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.471099] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 719.471099] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.471099] env[62753]: ERROR nova.compute.manager raise self.value [ 719.471099] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 719.471099] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 719.471099] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.471099] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 719.471612] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.471612] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 719.471612] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c2f33405-b9de-41af-9eed-37383de235f0, please check neutron logs for more information. [ 719.471612] env[62753]: ERROR nova.compute.manager [ 719.471612] env[62753]: Traceback (most recent call last): [ 719.471612] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 719.471612] env[62753]: listener.cb(fileno) [ 719.471612] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.471612] env[62753]: result = function(*args, **kwargs) [ 719.471612] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 719.471612] env[62753]: return func(*args, **kwargs) [ 719.471612] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 719.471612] env[62753]: raise e [ 719.471612] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.471612] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 719.471612] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 719.471612] env[62753]: created_port_ids = self._update_ports_for_instance( [ 719.471612] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 719.471612] env[62753]: with excutils.save_and_reraise_exception(): [ 719.471612] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.471612] env[62753]: self.force_reraise() [ 719.471612] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.471612] env[62753]: raise self.value [ 719.471612] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 719.471612] env[62753]: updated_port = self._update_port( [ 719.471612] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.471612] env[62753]: _ensure_no_port_binding_failure(port) [ 719.471612] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.471612] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 719.472349] env[62753]: nova.exception.PortBindingFailed: Binding failed for port c2f33405-b9de-41af-9eed-37383de235f0, please check neutron logs for more information. [ 719.472349] env[62753]: Removing descriptor: 14 [ 719.472349] env[62753]: ERROR nova.compute.manager [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c2f33405-b9de-41af-9eed-37383de235f0, please check neutron logs for more information. [ 719.472349] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Traceback (most recent call last): [ 719.472349] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 719.472349] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] yield resources [ 719.472349] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 719.472349] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] self.driver.spawn(context, instance, image_meta, [ 719.472349] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 719.472349] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] self._vmops.spawn(context, instance, image_meta, injected_files, [ 719.472349] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 719.472349] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] vm_ref = self.build_virtual_machine(instance, [ 719.472667] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 719.472667] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] vif_infos = vmwarevif.get_vif_info(self._session, [ 719.472667] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 719.472667] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] for vif in network_info: [ 719.472667] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 719.472667] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] return self._sync_wrapper(fn, *args, **kwargs) [ 719.472667] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 719.472667] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] self.wait() [ 719.472667] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 719.472667] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] self[:] = self._gt.wait() [ 719.472667] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 719.472667] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] return self._exit_event.wait() [ 719.472667] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 719.472989] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] result = hub.switch() [ 719.472989] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 719.472989] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] return self.greenlet.switch() [ 719.472989] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 719.472989] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] result = function(*args, **kwargs) [ 719.472989] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 719.472989] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] return func(*args, **kwargs) [ 719.472989] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 719.472989] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] raise e [ 719.472989] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 719.472989] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] nwinfo = self.network_api.allocate_for_instance( [ 719.472989] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 719.472989] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] created_port_ids = self._update_ports_for_instance( [ 719.473342] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 719.473342] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] with excutils.save_and_reraise_exception(): [ 719.473342] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 719.473342] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] self.force_reraise() [ 719.473342] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 719.473342] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] raise self.value [ 719.473342] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 719.473342] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] updated_port = self._update_port( [ 719.473342] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 719.473342] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] _ensure_no_port_binding_failure(port) [ 719.473342] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 719.473342] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] raise exception.PortBindingFailed(port_id=port['id']) [ 719.473668] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] nova.exception.PortBindingFailed: Binding failed for port c2f33405-b9de-41af-9eed-37383de235f0, please check neutron logs for more information. [ 719.473668] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] [ 719.473668] env[62753]: INFO nova.compute.manager [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Terminating instance [ 719.474399] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Acquiring lock "refresh_cache-50a480ed-f320-45d2-8ac9-a42d01614d11" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 719.474563] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Acquired lock "refresh_cache-50a480ed-f320-45d2-8ac9-a42d01614d11" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 719.474717] env[62753]: DEBUG nova.network.neutron [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 719.571767] env[62753]: DEBUG nova.network.neutron [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 719.645158] env[62753]: DEBUG nova.network.neutron [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 719.841624] env[62753]: INFO nova.compute.manager [-] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Took 1.03 seconds to deallocate network for instance. [ 719.844039] env[62753]: DEBUG nova.compute.claims [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 719.844228] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 719.990836] env[62753]: DEBUG nova.network.neutron [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.055119] env[62753]: DEBUG nova.network.neutron [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.147869] env[62753]: DEBUG oslo_concurrency.lockutils [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Releasing lock "refresh_cache-f10ff402-4308-4c75-af79-26e2942e0b92" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.148110] env[62753]: DEBUG nova.compute.manager [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 720.148356] env[62753]: DEBUG nova.compute.manager [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.148456] env[62753]: DEBUG nova.network.neutron [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 720.164522] env[62753]: DEBUG nova.network.neutron [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.370651] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f26ddd6-9216-40eb-ae01-a2bc6ef40ee6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.377993] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c45318e-687a-4291-9703-705fe3bfb09d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.407808] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7691397-ad14-4ba9-8882-92d11341d8dc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.415138] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57d40df8-b576-4ead-91b0-b01b22265e8e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.429514] env[62753]: DEBUG nova.compute.provider_tree [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 720.561877] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Releasing lock "refresh_cache-50a480ed-f320-45d2-8ac9-a42d01614d11" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 720.561877] env[62753]: DEBUG nova.compute.manager [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 720.561877] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 720.561877] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a3e1aa51-1075-4dac-96bf-9282fa8b3b49 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.570453] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7ef6541-2476-4731-81f5-020f620877d7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 720.591498] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 50a480ed-f320-45d2-8ac9-a42d01614d11 could not be found. [ 720.591727] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 720.591907] env[62753]: INFO nova.compute.manager [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Took 0.03 seconds to destroy the instance on the hypervisor. [ 720.592161] env[62753]: DEBUG oslo.service.loopingcall [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 720.593391] env[62753]: DEBUG nova.compute.manager [-] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 720.593391] env[62753]: DEBUG nova.network.neutron [-] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 720.595722] env[62753]: DEBUG nova.compute.manager [req-f478b8ae-bd2d-47f5-a655-4071875ccbb9 req-b0ddd314-c46a-4c29-a563-33cbf23010cd service nova] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Received event network-changed-c2f33405-b9de-41af-9eed-37383de235f0 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 720.595894] env[62753]: DEBUG nova.compute.manager [req-f478b8ae-bd2d-47f5-a655-4071875ccbb9 req-b0ddd314-c46a-4c29-a563-33cbf23010cd service nova] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Refreshing instance network info cache due to event network-changed-c2f33405-b9de-41af-9eed-37383de235f0. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 720.596114] env[62753]: DEBUG oslo_concurrency.lockutils [req-f478b8ae-bd2d-47f5-a655-4071875ccbb9 req-b0ddd314-c46a-4c29-a563-33cbf23010cd service nova] Acquiring lock "refresh_cache-50a480ed-f320-45d2-8ac9-a42d01614d11" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 720.596257] env[62753]: DEBUG oslo_concurrency.lockutils [req-f478b8ae-bd2d-47f5-a655-4071875ccbb9 req-b0ddd314-c46a-4c29-a563-33cbf23010cd service nova] Acquired lock "refresh_cache-50a480ed-f320-45d2-8ac9-a42d01614d11" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 720.596413] env[62753]: DEBUG nova.network.neutron [req-f478b8ae-bd2d-47f5-a655-4071875ccbb9 req-b0ddd314-c46a-4c29-a563-33cbf23010cd service nova] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Refreshing network info cache for port c2f33405-b9de-41af-9eed-37383de235f0 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 720.615426] env[62753]: DEBUG nova.network.neutron [-] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 720.666934] env[62753]: DEBUG nova.network.neutron [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 720.932959] env[62753]: DEBUG nova.scheduler.client.report [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 721.111677] env[62753]: DEBUG nova.network.neutron [req-f478b8ae-bd2d-47f5-a655-4071875ccbb9 req-b0ddd314-c46a-4c29-a563-33cbf23010cd service nova] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 721.117352] env[62753]: DEBUG nova.network.neutron [-] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.154405] env[62753]: DEBUG nova.network.neutron [req-f478b8ae-bd2d-47f5-a655-4071875ccbb9 req-b0ddd314-c46a-4c29-a563-33cbf23010cd service nova] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 721.169521] env[62753]: INFO nova.compute.manager [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: f10ff402-4308-4c75-af79-26e2942e0b92] Took 1.02 seconds to deallocate network for instance. [ 721.438849] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.394s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 721.438849] env[62753]: DEBUG nova.compute.manager [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 721.442026] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 30.658s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 721.443486] env[62753]: INFO nova.compute.claims [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 721.620391] env[62753]: INFO nova.compute.manager [-] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Took 1.03 seconds to deallocate network for instance. [ 721.622093] env[62753]: DEBUG nova.compute.claims [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 721.622823] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 721.656416] env[62753]: DEBUG oslo_concurrency.lockutils [req-f478b8ae-bd2d-47f5-a655-4071875ccbb9 req-b0ddd314-c46a-4c29-a563-33cbf23010cd service nova] Releasing lock "refresh_cache-50a480ed-f320-45d2-8ac9-a42d01614d11" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 721.656640] env[62753]: DEBUG nova.compute.manager [req-f478b8ae-bd2d-47f5-a655-4071875ccbb9 req-b0ddd314-c46a-4c29-a563-33cbf23010cd service nova] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Received event network-vif-deleted-c2f33405-b9de-41af-9eed-37383de235f0 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 721.949846] env[62753]: DEBUG nova.compute.utils [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 721.954040] env[62753]: DEBUG nova.compute.manager [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 721.954040] env[62753]: DEBUG nova.network.neutron [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 721.999986] env[62753]: DEBUG nova.policy [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e32e693cb471498494e211632025e302', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0549397fd54c4470a4dee75dc5e37547', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 722.198782] env[62753]: INFO nova.scheduler.client.report [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Deleted allocations for instance f10ff402-4308-4c75-af79-26e2942e0b92 [ 722.311119] env[62753]: DEBUG nova.network.neutron [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Successfully created port: c8a4dec9-bd48-48eb-900d-3dee1f1aa326 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 722.453845] env[62753]: DEBUG nova.compute.manager [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 722.711376] env[62753]: DEBUG oslo_concurrency.lockutils [None req-df95d4f9-d6e7-435b-a83c-196334332633 tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Lock "f10ff402-4308-4c75-af79-26e2942e0b92" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.776s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 722.864226] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a5fc2a4-6bca-4da8-94d1-7d7b3c2231f5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.877910] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6fb8d45-f742-4751-a7f8-ca4f5b8b48aa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.911887] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7200f531-457a-4ea8-9364-574cb8ff697d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.921283] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc373bc3-f7c5-41cf-9214-6bd850521dd3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 722.937939] env[62753]: DEBUG nova.compute.provider_tree [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 723.215918] env[62753]: DEBUG nova.compute.manager [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 723.441223] env[62753]: DEBUG nova.scheduler.client.report [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 723.470353] env[62753]: DEBUG nova.compute.manager [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 723.494829] env[62753]: DEBUG nova.virt.hardware [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 723.495273] env[62753]: DEBUG nova.virt.hardware [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 723.495497] env[62753]: DEBUG nova.virt.hardware [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 723.495731] env[62753]: DEBUG nova.virt.hardware [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 723.495884] env[62753]: DEBUG nova.virt.hardware [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 723.496084] env[62753]: DEBUG nova.virt.hardware [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 723.496316] env[62753]: DEBUG nova.virt.hardware [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 723.496479] env[62753]: DEBUG nova.virt.hardware [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 723.496644] env[62753]: DEBUG nova.virt.hardware [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 723.496806] env[62753]: DEBUG nova.virt.hardware [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 723.496975] env[62753]: DEBUG nova.virt.hardware [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 723.497835] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13baa679-6331-447a-bb35-82ab32bf6e73 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.505523] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-409093c1-3733-4bcc-93cf-04953e15eb03 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 723.656093] env[62753]: DEBUG nova.compute.manager [req-9bca30c1-0baa-45ea-83a6-8aa275bf7dbc req-edac9c2a-9fa2-4eb7-8efb-c5a1d8ec1c7f service nova] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Received event network-changed-c8a4dec9-bd48-48eb-900d-3dee1f1aa326 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 723.656263] env[62753]: DEBUG nova.compute.manager [req-9bca30c1-0baa-45ea-83a6-8aa275bf7dbc req-edac9c2a-9fa2-4eb7-8efb-c5a1d8ec1c7f service nova] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Refreshing instance network info cache due to event network-changed-c8a4dec9-bd48-48eb-900d-3dee1f1aa326. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 723.656583] env[62753]: DEBUG oslo_concurrency.lockutils [req-9bca30c1-0baa-45ea-83a6-8aa275bf7dbc req-edac9c2a-9fa2-4eb7-8efb-c5a1d8ec1c7f service nova] Acquiring lock "refresh_cache-250f2711-6f32-4c4d-9876-35a3a1c30d12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.656720] env[62753]: DEBUG oslo_concurrency.lockutils [req-9bca30c1-0baa-45ea-83a6-8aa275bf7dbc req-edac9c2a-9fa2-4eb7-8efb-c5a1d8ec1c7f service nova] Acquired lock "refresh_cache-250f2711-6f32-4c4d-9876-35a3a1c30d12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 723.656880] env[62753]: DEBUG nova.network.neutron [req-9bca30c1-0baa-45ea-83a6-8aa275bf7dbc req-edac9c2a-9fa2-4eb7-8efb-c5a1d8ec1c7f service nova] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Refreshing network info cache for port c8a4dec9-bd48-48eb-900d-3dee1f1aa326 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 723.747788] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 723.910415] env[62753]: ERROR nova.compute.manager [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port c8a4dec9-bd48-48eb-900d-3dee1f1aa326, please check neutron logs for more information. [ 723.910415] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 723.910415] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.910415] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 723.910415] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.910415] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 723.910415] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.910415] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 723.910415] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.910415] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 723.910415] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.910415] env[62753]: ERROR nova.compute.manager raise self.value [ 723.910415] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.910415] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 723.910415] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.910415] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 723.910874] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.910874] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 723.910874] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port c8a4dec9-bd48-48eb-900d-3dee1f1aa326, please check neutron logs for more information. [ 723.910874] env[62753]: ERROR nova.compute.manager [ 723.910874] env[62753]: Traceback (most recent call last): [ 723.910874] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 723.910874] env[62753]: listener.cb(fileno) [ 723.910874] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.910874] env[62753]: result = function(*args, **kwargs) [ 723.910874] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 723.910874] env[62753]: return func(*args, **kwargs) [ 723.910874] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.910874] env[62753]: raise e [ 723.910874] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.910874] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 723.910874] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.910874] env[62753]: created_port_ids = self._update_ports_for_instance( [ 723.910874] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.910874] env[62753]: with excutils.save_and_reraise_exception(): [ 723.910874] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.910874] env[62753]: self.force_reraise() [ 723.910874] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.910874] env[62753]: raise self.value [ 723.910874] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.910874] env[62753]: updated_port = self._update_port( [ 723.910874] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.910874] env[62753]: _ensure_no_port_binding_failure(port) [ 723.910874] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.910874] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 723.911631] env[62753]: nova.exception.PortBindingFailed: Binding failed for port c8a4dec9-bd48-48eb-900d-3dee1f1aa326, please check neutron logs for more information. [ 723.911631] env[62753]: Removing descriptor: 14 [ 723.911983] env[62753]: ERROR nova.compute.manager [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port c8a4dec9-bd48-48eb-900d-3dee1f1aa326, please check neutron logs for more information. [ 723.911983] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Traceback (most recent call last): [ 723.911983] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 723.911983] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] yield resources [ 723.911983] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 723.911983] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] self.driver.spawn(context, instance, image_meta, [ 723.911983] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 723.911983] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] self._vmops.spawn(context, instance, image_meta, injected_files, [ 723.911983] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 723.911983] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] vm_ref = self.build_virtual_machine(instance, [ 723.911983] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 723.912514] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] vif_infos = vmwarevif.get_vif_info(self._session, [ 723.912514] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 723.912514] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] for vif in network_info: [ 723.912514] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 723.912514] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] return self._sync_wrapper(fn, *args, **kwargs) [ 723.912514] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 723.912514] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] self.wait() [ 723.912514] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 723.912514] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] self[:] = self._gt.wait() [ 723.912514] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 723.912514] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] return self._exit_event.wait() [ 723.912514] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 723.912514] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] result = hub.switch() [ 723.912837] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 723.912837] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] return self.greenlet.switch() [ 723.912837] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 723.912837] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] result = function(*args, **kwargs) [ 723.912837] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 723.912837] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] return func(*args, **kwargs) [ 723.912837] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 723.912837] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] raise e [ 723.912837] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 723.912837] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] nwinfo = self.network_api.allocate_for_instance( [ 723.912837] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 723.912837] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] created_port_ids = self._update_ports_for_instance( [ 723.912837] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 723.913341] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] with excutils.save_and_reraise_exception(): [ 723.913341] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 723.913341] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] self.force_reraise() [ 723.913341] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 723.913341] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] raise self.value [ 723.913341] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 723.913341] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] updated_port = self._update_port( [ 723.913341] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 723.913341] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] _ensure_no_port_binding_failure(port) [ 723.913341] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 723.913341] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] raise exception.PortBindingFailed(port_id=port['id']) [ 723.913341] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] nova.exception.PortBindingFailed: Binding failed for port c8a4dec9-bd48-48eb-900d-3dee1f1aa326, please check neutron logs for more information. [ 723.913341] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] [ 723.913787] env[62753]: INFO nova.compute.manager [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Terminating instance [ 723.914297] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Acquiring lock "refresh_cache-250f2711-6f32-4c4d-9876-35a3a1c30d12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 723.946237] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.504s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 723.947024] env[62753]: DEBUG nova.compute.manager [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 723.949543] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 29.586s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 723.951060] env[62753]: INFO nova.compute.claims [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 724.183746] env[62753]: DEBUG nova.network.neutron [req-9bca30c1-0baa-45ea-83a6-8aa275bf7dbc req-edac9c2a-9fa2-4eb7-8efb-c5a1d8ec1c7f service nova] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 724.303171] env[62753]: DEBUG nova.network.neutron [req-9bca30c1-0baa-45ea-83a6-8aa275bf7dbc req-edac9c2a-9fa2-4eb7-8efb-c5a1d8ec1c7f service nova] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 724.455327] env[62753]: DEBUG nova.compute.utils [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 724.458305] env[62753]: DEBUG nova.compute.manager [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 724.458475] env[62753]: DEBUG nova.network.neutron [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 724.527281] env[62753]: DEBUG nova.policy [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4d2a40cf18254afbbcda56d7686d5702', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b0221d2c229640b1925bb14f83d4118b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 724.805705] env[62753]: DEBUG oslo_concurrency.lockutils [req-9bca30c1-0baa-45ea-83a6-8aa275bf7dbc req-edac9c2a-9fa2-4eb7-8efb-c5a1d8ec1c7f service nova] Releasing lock "refresh_cache-250f2711-6f32-4c4d-9876-35a3a1c30d12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 724.806146] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Acquired lock "refresh_cache-250f2711-6f32-4c4d-9876-35a3a1c30d12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 724.806409] env[62753]: DEBUG nova.network.neutron [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 724.961299] env[62753]: DEBUG nova.compute.manager [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 725.062653] env[62753]: DEBUG nova.network.neutron [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Successfully created port: db911c7c-0d38-4274-b8c2-1c40aa502460 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 725.323502] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06729799-4ca3-45c3-9c0b-4d7c7579562b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.325348] env[62753]: DEBUG nova.network.neutron [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.331556] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d2a116e-fc89-4a50-86c3-314b639d45e8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.360719] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc59b762-d6b9-4bab-acc4-6070643656f0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.370795] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e6d94bd-a031-41f1-918d-83574ed1058b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.384046] env[62753]: DEBUG nova.compute.provider_tree [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 725.410167] env[62753]: DEBUG nova.network.neutron [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 725.789822] env[62753]: DEBUG nova.compute.manager [req-240bc87e-bbe9-45cf-a3cb-4bdb38a112fd req-bc6f2285-d6b8-4b4b-9d77-6abb5eaeaaf1 service nova] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Received event network-vif-deleted-c8a4dec9-bd48-48eb-900d-3dee1f1aa326 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 725.887087] env[62753]: DEBUG nova.scheduler.client.report [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 725.912616] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Releasing lock "refresh_cache-250f2711-6f32-4c4d-9876-35a3a1c30d12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 725.913013] env[62753]: DEBUG nova.compute.manager [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 725.913222] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 725.913886] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-94b4168f-d4b7-47d1-ab5c-e8ae10fbc408 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.923015] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82307adb-2821-4ee8-9edb-5c6ce99ec925 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 725.951878] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 250f2711-6f32-4c4d-9876-35a3a1c30d12 could not be found. [ 725.952367] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 725.952607] env[62753]: INFO nova.compute.manager [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Took 0.04 seconds to destroy the instance on the hypervisor. [ 725.952828] env[62753]: DEBUG oslo.service.loopingcall [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 725.953065] env[62753]: DEBUG nova.compute.manager [-] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 725.953163] env[62753]: DEBUG nova.network.neutron [-] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 725.969170] env[62753]: DEBUG nova.network.neutron [-] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 725.973908] env[62753]: DEBUG nova.compute.manager [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 725.999852] env[62753]: DEBUG nova.virt.hardware [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 726.000124] env[62753]: DEBUG nova.virt.hardware [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 726.000256] env[62753]: DEBUG nova.virt.hardware [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 726.000438] env[62753]: DEBUG nova.virt.hardware [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 726.000582] env[62753]: DEBUG nova.virt.hardware [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 726.000729] env[62753]: DEBUG nova.virt.hardware [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 726.000935] env[62753]: DEBUG nova.virt.hardware [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 726.001345] env[62753]: DEBUG nova.virt.hardware [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 726.001457] env[62753]: DEBUG nova.virt.hardware [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 726.001601] env[62753]: DEBUG nova.virt.hardware [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 726.001771] env[62753]: DEBUG nova.virt.hardware [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 726.002614] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1adbc374-f2c4-4fa1-a360-a6d408155b44 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.010271] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6db87cc9-a57f-43fe-8493-63a1a612a96d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 726.396518] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.447s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 726.397308] env[62753]: DEBUG nova.compute.manager [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 726.399569] env[62753]: DEBUG oslo_concurrency.lockutils [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 30.597s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 726.471908] env[62753]: DEBUG nova.network.neutron [-] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 726.482260] env[62753]: ERROR nova.compute.manager [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port db911c7c-0d38-4274-b8c2-1c40aa502460, please check neutron logs for more information. [ 726.482260] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 726.482260] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.482260] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 726.482260] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.482260] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 726.482260] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.482260] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 726.482260] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.482260] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 726.482260] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.482260] env[62753]: ERROR nova.compute.manager raise self.value [ 726.482260] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.482260] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 726.482260] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.482260] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 726.482724] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.482724] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 726.482724] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port db911c7c-0d38-4274-b8c2-1c40aa502460, please check neutron logs for more information. [ 726.482724] env[62753]: ERROR nova.compute.manager [ 726.482724] env[62753]: Traceback (most recent call last): [ 726.482724] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 726.482724] env[62753]: listener.cb(fileno) [ 726.482724] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.482724] env[62753]: result = function(*args, **kwargs) [ 726.482724] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 726.482724] env[62753]: return func(*args, **kwargs) [ 726.482724] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.482724] env[62753]: raise e [ 726.482724] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.482724] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 726.482724] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.482724] env[62753]: created_port_ids = self._update_ports_for_instance( [ 726.482724] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.482724] env[62753]: with excutils.save_and_reraise_exception(): [ 726.482724] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.482724] env[62753]: self.force_reraise() [ 726.482724] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.482724] env[62753]: raise self.value [ 726.482724] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.482724] env[62753]: updated_port = self._update_port( [ 726.482724] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.482724] env[62753]: _ensure_no_port_binding_failure(port) [ 726.482724] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.482724] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 726.483431] env[62753]: nova.exception.PortBindingFailed: Binding failed for port db911c7c-0d38-4274-b8c2-1c40aa502460, please check neutron logs for more information. [ 726.483431] env[62753]: Removing descriptor: 14 [ 726.483431] env[62753]: ERROR nova.compute.manager [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port db911c7c-0d38-4274-b8c2-1c40aa502460, please check neutron logs for more information. [ 726.483431] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Traceback (most recent call last): [ 726.483431] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 726.483431] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] yield resources [ 726.483431] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 726.483431] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] self.driver.spawn(context, instance, image_meta, [ 726.483431] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 726.483431] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 726.483431] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 726.483431] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] vm_ref = self.build_virtual_machine(instance, [ 726.483735] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 726.483735] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 726.483735] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 726.483735] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] for vif in network_info: [ 726.483735] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 726.483735] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] return self._sync_wrapper(fn, *args, **kwargs) [ 726.483735] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 726.483735] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] self.wait() [ 726.483735] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 726.483735] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] self[:] = self._gt.wait() [ 726.483735] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 726.483735] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] return self._exit_event.wait() [ 726.483735] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 726.484071] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] result = hub.switch() [ 726.484071] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 726.484071] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] return self.greenlet.switch() [ 726.484071] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 726.484071] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] result = function(*args, **kwargs) [ 726.484071] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 726.484071] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] return func(*args, **kwargs) [ 726.484071] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 726.484071] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] raise e [ 726.484071] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 726.484071] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] nwinfo = self.network_api.allocate_for_instance( [ 726.484071] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 726.484071] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] created_port_ids = self._update_ports_for_instance( [ 726.484448] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 726.484448] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] with excutils.save_and_reraise_exception(): [ 726.484448] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 726.484448] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] self.force_reraise() [ 726.484448] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 726.484448] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] raise self.value [ 726.484448] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 726.484448] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] updated_port = self._update_port( [ 726.484448] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 726.484448] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] _ensure_no_port_binding_failure(port) [ 726.484448] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 726.484448] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] raise exception.PortBindingFailed(port_id=port['id']) [ 726.484735] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] nova.exception.PortBindingFailed: Binding failed for port db911c7c-0d38-4274-b8c2-1c40aa502460, please check neutron logs for more information. [ 726.484735] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] [ 726.484735] env[62753]: INFO nova.compute.manager [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Terminating instance [ 726.485346] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Acquiring lock "refresh_cache-a413174f-2a97-4c41-b512-e8ab562a25f4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 726.485541] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Acquired lock "refresh_cache-a413174f-2a97-4c41-b512-e8ab562a25f4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 726.485719] env[62753]: DEBUG nova.network.neutron [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 726.907871] env[62753]: DEBUG nova.compute.utils [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 726.909611] env[62753]: DEBUG nova.compute.manager [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 726.909611] env[62753]: DEBUG nova.network.neutron [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 726.974668] env[62753]: INFO nova.compute.manager [-] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Took 1.02 seconds to deallocate network for instance. [ 726.977295] env[62753]: DEBUG nova.compute.claims [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 726.977725] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 726.979766] env[62753]: DEBUG nova.policy [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cbc9dc744e144ff28cf4c44fb043094f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e376836411ae4854965adf2923eaaffd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 727.008888] env[62753]: DEBUG nova.network.neutron [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.091605] env[62753]: DEBUG nova.network.neutron [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 727.234803] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8df81c6-e8fa-429f-9383-b174971f5bda {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.243008] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65746072-d7fa-4a8f-bcc3-80065f6fa305 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.278027] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b6bba6-df49-4aec-a4f9-e461dcd3ff28 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.284948] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-101b4eb9-4e03-4577-b0d9-2958ce3eecc1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.299595] env[62753]: DEBUG nova.compute.provider_tree [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 727.401841] env[62753]: DEBUG nova.network.neutron [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Successfully created port: 8b4fcaab-c1f4-4796-9d86-8de53459855a {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 727.413034] env[62753]: DEBUG nova.compute.manager [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 727.594164] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Releasing lock "refresh_cache-a413174f-2a97-4c41-b512-e8ab562a25f4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 727.594563] env[62753]: DEBUG nova.compute.manager [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 727.594759] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 727.595071] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d94557e9-9537-4d6a-9e51-27b208ccac31 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.604514] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-957adda5-9661-4ae2-8000-3c39b6c4becc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 727.626193] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance a413174f-2a97-4c41-b512-e8ab562a25f4 could not be found. [ 727.626417] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 727.626598] env[62753]: INFO nova.compute.manager [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Took 0.03 seconds to destroy the instance on the hypervisor. [ 727.626838] env[62753]: DEBUG oslo.service.loopingcall [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 727.627084] env[62753]: DEBUG nova.compute.manager [-] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 727.627181] env[62753]: DEBUG nova.network.neutron [-] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 727.646378] env[62753]: DEBUG nova.network.neutron [-] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 727.803036] env[62753]: DEBUG nova.scheduler.client.report [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 727.897408] env[62753]: DEBUG nova.compute.manager [req-1e34e340-11ef-40fa-8e42-d48592337b84 req-30400b78-3afc-4f6f-ad30-6a33cf963531 service nova] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Received event network-changed-db911c7c-0d38-4274-b8c2-1c40aa502460 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 727.897408] env[62753]: DEBUG nova.compute.manager [req-1e34e340-11ef-40fa-8e42-d48592337b84 req-30400b78-3afc-4f6f-ad30-6a33cf963531 service nova] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Refreshing instance network info cache due to event network-changed-db911c7c-0d38-4274-b8c2-1c40aa502460. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 727.897408] env[62753]: DEBUG oslo_concurrency.lockutils [req-1e34e340-11ef-40fa-8e42-d48592337b84 req-30400b78-3afc-4f6f-ad30-6a33cf963531 service nova] Acquiring lock "refresh_cache-a413174f-2a97-4c41-b512-e8ab562a25f4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 727.897408] env[62753]: DEBUG oslo_concurrency.lockutils [req-1e34e340-11ef-40fa-8e42-d48592337b84 req-30400b78-3afc-4f6f-ad30-6a33cf963531 service nova] Acquired lock "refresh_cache-a413174f-2a97-4c41-b512-e8ab562a25f4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 727.897408] env[62753]: DEBUG nova.network.neutron [req-1e34e340-11ef-40fa-8e42-d48592337b84 req-30400b78-3afc-4f6f-ad30-6a33cf963531 service nova] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Refreshing network info cache for port db911c7c-0d38-4274-b8c2-1c40aa502460 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 727.964893] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Acquiring lock "7931ae4b-c486-4e57-9bf5-14166a1d4201" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 727.965149] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Lock "7931ae4b-c486-4e57-9bf5-14166a1d4201" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.148383] env[62753]: DEBUG nova.network.neutron [-] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.308355] env[62753]: DEBUG oslo_concurrency.lockutils [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.909s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 728.309071] env[62753]: ERROR nova.compute.manager [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port e0e7e24b-752f-4598-b03e-3ac3e1ec3141, please check neutron logs for more information. [ 728.309071] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Traceback (most recent call last): [ 728.309071] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 728.309071] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] self.driver.spawn(context, instance, image_meta, [ 728.309071] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 728.309071] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.309071] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.309071] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] vm_ref = self.build_virtual_machine(instance, [ 728.309071] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.309071] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.309071] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.309672] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] for vif in network_info: [ 728.309672] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.309672] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] return self._sync_wrapper(fn, *args, **kwargs) [ 728.309672] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.309672] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] self.wait() [ 728.309672] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.309672] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] self[:] = self._gt.wait() [ 728.309672] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.309672] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] return self._exit_event.wait() [ 728.309672] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 728.309672] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] current.throw(*self._exc) [ 728.309672] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.309672] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] result = function(*args, **kwargs) [ 728.310225] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 728.310225] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] return func(*args, **kwargs) [ 728.310225] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.310225] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] raise e [ 728.310225] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.310225] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] nwinfo = self.network_api.allocate_for_instance( [ 728.310225] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.310225] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] created_port_ids = self._update_ports_for_instance( [ 728.310225] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.310225] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] with excutils.save_and_reraise_exception(): [ 728.310225] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.310225] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] self.force_reraise() [ 728.310225] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.310739] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] raise self.value [ 728.310739] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.310739] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] updated_port = self._update_port( [ 728.310739] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.310739] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] _ensure_no_port_binding_failure(port) [ 728.310739] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.310739] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] raise exception.PortBindingFailed(port_id=port['id']) [ 728.310739] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] nova.exception.PortBindingFailed: Binding failed for port e0e7e24b-752f-4598-b03e-3ac3e1ec3141, please check neutron logs for more information. [ 728.310739] env[62753]: ERROR nova.compute.manager [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] [ 728.310739] env[62753]: DEBUG nova.compute.utils [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Binding failed for port e0e7e24b-752f-4598-b03e-3ac3e1ec3141, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 728.311487] env[62753]: DEBUG oslo_concurrency.lockutils [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.761s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 728.314389] env[62753]: DEBUG nova.compute.manager [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Build of instance 5fcbd37f-0568-43db-ae44-596eb4467391 was re-scheduled: Binding failed for port e0e7e24b-752f-4598-b03e-3ac3e1ec3141, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 728.314816] env[62753]: DEBUG nova.compute.manager [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 728.315054] env[62753]: DEBUG oslo_concurrency.lockutils [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Acquiring lock "refresh_cache-5fcbd37f-0568-43db-ae44-596eb4467391" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.315204] env[62753]: DEBUG oslo_concurrency.lockutils [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Acquired lock "refresh_cache-5fcbd37f-0568-43db-ae44-596eb4467391" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.315361] env[62753]: DEBUG nova.network.neutron [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 728.419697] env[62753]: DEBUG nova.network.neutron [req-1e34e340-11ef-40fa-8e42-d48592337b84 req-30400b78-3afc-4f6f-ad30-6a33cf963531 service nova] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.425461] env[62753]: DEBUG nova.compute.manager [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 728.458831] env[62753]: DEBUG nova.virt.hardware [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 728.461949] env[62753]: DEBUG nova.virt.hardware [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 728.461949] env[62753]: DEBUG nova.virt.hardware [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 728.461949] env[62753]: DEBUG nova.virt.hardware [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 728.461949] env[62753]: DEBUG nova.virt.hardware [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 728.461949] env[62753]: DEBUG nova.virt.hardware [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 728.461949] env[62753]: DEBUG nova.virt.hardware [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 728.462209] env[62753]: DEBUG nova.virt.hardware [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 728.462209] env[62753]: DEBUG nova.virt.hardware [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 728.462209] env[62753]: DEBUG nova.virt.hardware [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 728.462209] env[62753]: DEBUG nova.virt.hardware [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 728.462894] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55e97d9a-40dd-4018-a988-bf255b76842c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.473886] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e808afe-03ac-4d6a-b2e6-24d62f882570 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 728.507520] env[62753]: DEBUG nova.network.neutron [req-1e34e340-11ef-40fa-8e42-d48592337b84 req-30400b78-3afc-4f6f-ad30-6a33cf963531 service nova] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 728.651123] env[62753]: INFO nova.compute.manager [-] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Took 1.02 seconds to deallocate network for instance. [ 728.655092] env[62753]: DEBUG nova.compute.claims [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 728.655276] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 728.790129] env[62753]: ERROR nova.compute.manager [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 8b4fcaab-c1f4-4796-9d86-8de53459855a, please check neutron logs for more information. [ 728.790129] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 728.790129] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.790129] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 728.790129] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.790129] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 728.790129] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.790129] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 728.790129] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.790129] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 728.790129] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.790129] env[62753]: ERROR nova.compute.manager raise self.value [ 728.790129] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.790129] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 728.790129] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.790129] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 728.790853] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.790853] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 728.790853] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 8b4fcaab-c1f4-4796-9d86-8de53459855a, please check neutron logs for more information. [ 728.790853] env[62753]: ERROR nova.compute.manager [ 728.790853] env[62753]: Traceback (most recent call last): [ 728.790853] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 728.790853] env[62753]: listener.cb(fileno) [ 728.790853] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.790853] env[62753]: result = function(*args, **kwargs) [ 728.790853] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 728.790853] env[62753]: return func(*args, **kwargs) [ 728.790853] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.790853] env[62753]: raise e [ 728.790853] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.790853] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 728.790853] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.790853] env[62753]: created_port_ids = self._update_ports_for_instance( [ 728.790853] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.790853] env[62753]: with excutils.save_and_reraise_exception(): [ 728.790853] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.790853] env[62753]: self.force_reraise() [ 728.790853] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.790853] env[62753]: raise self.value [ 728.790853] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.790853] env[62753]: updated_port = self._update_port( [ 728.790853] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.790853] env[62753]: _ensure_no_port_binding_failure(port) [ 728.790853] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.790853] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 728.791616] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 8b4fcaab-c1f4-4796-9d86-8de53459855a, please check neutron logs for more information. [ 728.791616] env[62753]: Removing descriptor: 14 [ 728.791616] env[62753]: ERROR nova.compute.manager [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 8b4fcaab-c1f4-4796-9d86-8de53459855a, please check neutron logs for more information. [ 728.791616] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Traceback (most recent call last): [ 728.791616] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 728.791616] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] yield resources [ 728.791616] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 728.791616] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] self.driver.spawn(context, instance, image_meta, [ 728.791616] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 728.791616] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 728.791616] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 728.791616] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] vm_ref = self.build_virtual_machine(instance, [ 728.791946] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 728.791946] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] vif_infos = vmwarevif.get_vif_info(self._session, [ 728.791946] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 728.791946] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] for vif in network_info: [ 728.791946] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 728.791946] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] return self._sync_wrapper(fn, *args, **kwargs) [ 728.791946] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 728.791946] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] self.wait() [ 728.791946] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 728.791946] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] self[:] = self._gt.wait() [ 728.791946] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 728.791946] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] return self._exit_event.wait() [ 728.791946] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 728.792296] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] result = hub.switch() [ 728.792296] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 728.792296] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] return self.greenlet.switch() [ 728.792296] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 728.792296] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] result = function(*args, **kwargs) [ 728.792296] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 728.792296] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] return func(*args, **kwargs) [ 728.792296] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 728.792296] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] raise e [ 728.792296] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 728.792296] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] nwinfo = self.network_api.allocate_for_instance( [ 728.792296] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 728.792296] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] created_port_ids = self._update_ports_for_instance( [ 728.792634] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 728.792634] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] with excutils.save_and_reraise_exception(): [ 728.792634] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 728.792634] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] self.force_reraise() [ 728.792634] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 728.792634] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] raise self.value [ 728.792634] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 728.792634] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] updated_port = self._update_port( [ 728.792634] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 728.792634] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] _ensure_no_port_binding_failure(port) [ 728.792634] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 728.792634] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] raise exception.PortBindingFailed(port_id=port['id']) [ 728.793599] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] nova.exception.PortBindingFailed: Binding failed for port 8b4fcaab-c1f4-4796-9d86-8de53459855a, please check neutron logs for more information. [ 728.793599] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] [ 728.793599] env[62753]: INFO nova.compute.manager [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Terminating instance [ 728.793599] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "refresh_cache-c4ccc036-4963-4499-8bbb-dad5c0e4687e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 728.793720] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired lock "refresh_cache-c4ccc036-4963-4499-8bbb-dad5c0e4687e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 728.793826] env[62753]: DEBUG nova.network.neutron [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 728.837032] env[62753]: DEBUG nova.network.neutron [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 728.970471] env[62753]: DEBUG nova.network.neutron [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.011053] env[62753]: DEBUG oslo_concurrency.lockutils [req-1e34e340-11ef-40fa-8e42-d48592337b84 req-30400b78-3afc-4f6f-ad30-6a33cf963531 service nova] Releasing lock "refresh_cache-a413174f-2a97-4c41-b512-e8ab562a25f4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.011339] env[62753]: DEBUG nova.compute.manager [req-1e34e340-11ef-40fa-8e42-d48592337b84 req-30400b78-3afc-4f6f-ad30-6a33cf963531 service nova] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Received event network-vif-deleted-db911c7c-0d38-4274-b8c2-1c40aa502460 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 729.198046] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e12e5f-43e1-4f87-b0de-399a707ab6af {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.205604] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a287bf8-c26f-4d3d-b9b5-7cb4ff0792bb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.235329] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-338ced80-6423-476a-b88e-a037f7d6ea70 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.242432] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d6afa77-a925-4af4-a5e5-22489510a853 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.256365] env[62753]: DEBUG nova.compute.provider_tree [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 729.311388] env[62753]: DEBUG nova.network.neutron [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.378590] env[62753]: DEBUG nova.network.neutron [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 729.474898] env[62753]: DEBUG oslo_concurrency.lockutils [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Releasing lock "refresh_cache-5fcbd37f-0568-43db-ae44-596eb4467391" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.475160] env[62753]: DEBUG nova.compute.manager [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 729.475329] env[62753]: DEBUG nova.compute.manager [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 729.475496] env[62753]: DEBUG nova.network.neutron [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 729.496043] env[62753]: DEBUG nova.network.neutron [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.760884] env[62753]: DEBUG nova.scheduler.client.report [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 729.885138] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Releasing lock "refresh_cache-c4ccc036-4963-4499-8bbb-dad5c0e4687e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 729.885138] env[62753]: DEBUG nova.compute.manager [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 729.885138] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 729.885138] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-77263589-4e0f-46e0-9201-3e5b8cf975b5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.894447] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b027df1-1e21-4c34-9b51-bd424bb89e5c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 729.915947] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance c4ccc036-4963-4499-8bbb-dad5c0e4687e could not be found. [ 729.916201] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 729.916387] env[62753]: INFO nova.compute.manager [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Took 0.03 seconds to destroy the instance on the hypervisor. [ 729.916633] env[62753]: DEBUG oslo.service.loopingcall [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 729.916847] env[62753]: DEBUG nova.compute.manager [-] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 729.916941] env[62753]: DEBUG nova.network.neutron [-] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 729.920709] env[62753]: DEBUG nova.compute.manager [req-b0f27d69-f39b-4fc3-8ec8-20bc0d58bb44 req-d80ad6be-6014-4923-8147-84f8fcd6e437 service nova] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Received event network-changed-8b4fcaab-c1f4-4796-9d86-8de53459855a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 729.920952] env[62753]: DEBUG nova.compute.manager [req-b0f27d69-f39b-4fc3-8ec8-20bc0d58bb44 req-d80ad6be-6014-4923-8147-84f8fcd6e437 service nova] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Refreshing instance network info cache due to event network-changed-8b4fcaab-c1f4-4796-9d86-8de53459855a. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 729.921113] env[62753]: DEBUG oslo_concurrency.lockutils [req-b0f27d69-f39b-4fc3-8ec8-20bc0d58bb44 req-d80ad6be-6014-4923-8147-84f8fcd6e437 service nova] Acquiring lock "refresh_cache-c4ccc036-4963-4499-8bbb-dad5c0e4687e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 729.921256] env[62753]: DEBUG oslo_concurrency.lockutils [req-b0f27d69-f39b-4fc3-8ec8-20bc0d58bb44 req-d80ad6be-6014-4923-8147-84f8fcd6e437 service nova] Acquired lock "refresh_cache-c4ccc036-4963-4499-8bbb-dad5c0e4687e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 729.921441] env[62753]: DEBUG nova.network.neutron [req-b0f27d69-f39b-4fc3-8ec8-20bc0d58bb44 req-d80ad6be-6014-4923-8147-84f8fcd6e437 service nova] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Refreshing network info cache for port 8b4fcaab-c1f4-4796-9d86-8de53459855a {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 729.935588] env[62753]: DEBUG nova.network.neutron [-] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 729.998790] env[62753]: DEBUG nova.network.neutron [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.268795] env[62753]: DEBUG oslo_concurrency.lockutils [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.958s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 730.269482] env[62753]: ERROR nova.compute.manager [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ff4576be-cfd3-4fd8-b156-f822a6bf1651, please check neutron logs for more information. [ 730.269482] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Traceback (most recent call last): [ 730.269482] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 730.269482] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] self.driver.spawn(context, instance, image_meta, [ 730.269482] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 730.269482] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 730.269482] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 730.269482] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] vm_ref = self.build_virtual_machine(instance, [ 730.269482] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 730.269482] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] vif_infos = vmwarevif.get_vif_info(self._session, [ 730.269482] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 730.269856] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] for vif in network_info: [ 730.269856] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 730.269856] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] return self._sync_wrapper(fn, *args, **kwargs) [ 730.269856] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 730.269856] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] self.wait() [ 730.269856] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 730.269856] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] self[:] = self._gt.wait() [ 730.269856] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 730.269856] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] return self._exit_event.wait() [ 730.269856] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 730.269856] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] current.throw(*self._exc) [ 730.269856] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 730.269856] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] result = function(*args, **kwargs) [ 730.270245] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 730.270245] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] return func(*args, **kwargs) [ 730.270245] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 730.270245] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] raise e [ 730.270245] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 730.270245] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] nwinfo = self.network_api.allocate_for_instance( [ 730.270245] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 730.270245] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] created_port_ids = self._update_ports_for_instance( [ 730.270245] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 730.270245] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] with excutils.save_and_reraise_exception(): [ 730.270245] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 730.270245] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] self.force_reraise() [ 730.270245] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 730.270626] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] raise self.value [ 730.270626] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 730.270626] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] updated_port = self._update_port( [ 730.270626] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 730.270626] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] _ensure_no_port_binding_failure(port) [ 730.270626] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 730.270626] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] raise exception.PortBindingFailed(port_id=port['id']) [ 730.270626] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] nova.exception.PortBindingFailed: Binding failed for port ff4576be-cfd3-4fd8-b156-f822a6bf1651, please check neutron logs for more information. [ 730.270626] env[62753]: ERROR nova.compute.manager [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] [ 730.270626] env[62753]: DEBUG nova.compute.utils [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Binding failed for port ff4576be-cfd3-4fd8-b156-f822a6bf1651, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 730.271804] env[62753]: DEBUG oslo_concurrency.lockutils [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.989s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 730.276161] env[62753]: DEBUG nova.compute.manager [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Build of instance efaf8c7c-3f38-4a18-97e8-099da390c7d8 was re-scheduled: Binding failed for port ff4576be-cfd3-4fd8-b156-f822a6bf1651, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 730.276819] env[62753]: DEBUG nova.compute.manager [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 730.277171] env[62753]: DEBUG oslo_concurrency.lockutils [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Acquiring lock "refresh_cache-efaf8c7c-3f38-4a18-97e8-099da390c7d8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 730.277424] env[62753]: DEBUG oslo_concurrency.lockutils [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Acquired lock "refresh_cache-efaf8c7c-3f38-4a18-97e8-099da390c7d8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 730.277699] env[62753]: DEBUG nova.network.neutron [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 730.438480] env[62753]: DEBUG nova.network.neutron [req-b0f27d69-f39b-4fc3-8ec8-20bc0d58bb44 req-d80ad6be-6014-4923-8147-84f8fcd6e437 service nova] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.440042] env[62753]: DEBUG nova.network.neutron [-] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.501381] env[62753]: INFO nova.compute.manager [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] [instance: 5fcbd37f-0568-43db-ae44-596eb4467391] Took 1.03 seconds to deallocate network for instance. [ 730.515611] env[62753]: DEBUG nova.network.neutron [req-b0f27d69-f39b-4fc3-8ec8-20bc0d58bb44 req-d80ad6be-6014-4923-8147-84f8fcd6e437 service nova] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.799659] env[62753]: DEBUG nova.network.neutron [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 730.919570] env[62753]: DEBUG nova.network.neutron [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 730.944870] env[62753]: INFO nova.compute.manager [-] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Took 1.03 seconds to deallocate network for instance. [ 730.950069] env[62753]: DEBUG nova.compute.claims [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 730.950260] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 731.017957] env[62753]: DEBUG oslo_concurrency.lockutils [req-b0f27d69-f39b-4fc3-8ec8-20bc0d58bb44 req-d80ad6be-6014-4923-8147-84f8fcd6e437 service nova] Releasing lock "refresh_cache-c4ccc036-4963-4499-8bbb-dad5c0e4687e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.018232] env[62753]: DEBUG nova.compute.manager [req-b0f27d69-f39b-4fc3-8ec8-20bc0d58bb44 req-d80ad6be-6014-4923-8147-84f8fcd6e437 service nova] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Received event network-vif-deleted-8b4fcaab-c1f4-4796-9d86-8de53459855a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 731.116966] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f080c07-8c5b-461d-9dfb-9a2db86d2c4e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.124933] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9387002d-3059-41fb-a890-783ca6e6261f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.157021] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43c8641c-56c4-4ce8-997d-a35309d47778 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.170016] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad29b3aa-75c8-4c61-86ef-ee5a36258234 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 731.179251] env[62753]: DEBUG nova.compute.provider_tree [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 731.422042] env[62753]: DEBUG oslo_concurrency.lockutils [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Releasing lock "refresh_cache-efaf8c7c-3f38-4a18-97e8-099da390c7d8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 731.422042] env[62753]: DEBUG nova.compute.manager [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 731.422204] env[62753]: DEBUG nova.compute.manager [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 731.422331] env[62753]: DEBUG nova.network.neutron [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 731.435989] env[62753]: DEBUG nova.network.neutron [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 731.535651] env[62753]: INFO nova.scheduler.client.report [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Deleted allocations for instance 5fcbd37f-0568-43db-ae44-596eb4467391 [ 731.682887] env[62753]: DEBUG nova.scheduler.client.report [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 731.939651] env[62753]: DEBUG nova.network.neutron [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 732.044158] env[62753]: DEBUG oslo_concurrency.lockutils [None req-500359b9-fd9a-4880-9084-06fb29cf9187 tempest-InstanceActionsV221TestJSON-2019455691 tempest-InstanceActionsV221TestJSON-2019455691-project-member] Lock "5fcbd37f-0568-43db-ae44-596eb4467391" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 112.117s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.188035] env[62753]: DEBUG oslo_concurrency.lockutils [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.915s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 732.188035] env[62753]: ERROR nova.compute.manager [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port ff6cde93-e89f-4cec-9dc3-50d57ab15e2f, please check neutron logs for more information. [ 732.188035] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Traceback (most recent call last): [ 732.188035] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 732.188035] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] self.driver.spawn(context, instance, image_meta, [ 732.188035] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 732.188035] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] self._vmops.spawn(context, instance, image_meta, injected_files, [ 732.188035] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 732.188035] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] vm_ref = self.build_virtual_machine(instance, [ 732.188422] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 732.188422] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] vif_infos = vmwarevif.get_vif_info(self._session, [ 732.188422] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 732.188422] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] for vif in network_info: [ 732.188422] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 732.188422] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] return self._sync_wrapper(fn, *args, **kwargs) [ 732.188422] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 732.188422] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] self.wait() [ 732.188422] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 732.188422] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] self[:] = self._gt.wait() [ 732.188422] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 732.188422] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] return self._exit_event.wait() [ 732.188422] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 732.188862] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] current.throw(*self._exc) [ 732.188862] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 732.188862] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] result = function(*args, **kwargs) [ 732.188862] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 732.188862] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] return func(*args, **kwargs) [ 732.188862] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 732.188862] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] raise e [ 732.188862] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 732.188862] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] nwinfo = self.network_api.allocate_for_instance( [ 732.188862] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 732.188862] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] created_port_ids = self._update_ports_for_instance( [ 732.188862] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 732.188862] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] with excutils.save_and_reraise_exception(): [ 732.189325] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 732.189325] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] self.force_reraise() [ 732.189325] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 732.189325] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] raise self.value [ 732.189325] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 732.189325] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] updated_port = self._update_port( [ 732.189325] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 732.189325] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] _ensure_no_port_binding_failure(port) [ 732.189325] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 732.189325] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] raise exception.PortBindingFailed(port_id=port['id']) [ 732.189325] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] nova.exception.PortBindingFailed: Binding failed for port ff6cde93-e89f-4cec-9dc3-50d57ab15e2f, please check neutron logs for more information. [ 732.189325] env[62753]: ERROR nova.compute.manager [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] [ 732.189836] env[62753]: DEBUG nova.compute.utils [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Binding failed for port ff6cde93-e89f-4cec-9dc3-50d57ab15e2f, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 732.190111] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.432s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 732.191580] env[62753]: INFO nova.compute.claims [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 732.194889] env[62753]: DEBUG nova.compute.manager [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Build of instance 3f46fa44-4410-4ca9-98d0-8efa4cf06d52 was re-scheduled: Binding failed for port ff6cde93-e89f-4cec-9dc3-50d57ab15e2f, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 732.196137] env[62753]: DEBUG nova.compute.manager [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 732.196137] env[62753]: DEBUG oslo_concurrency.lockutils [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "refresh_cache-3f46fa44-4410-4ca9-98d0-8efa4cf06d52" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 732.196137] env[62753]: DEBUG oslo_concurrency.lockutils [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "refresh_cache-3f46fa44-4410-4ca9-98d0-8efa4cf06d52" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 732.196137] env[62753]: DEBUG nova.network.neutron [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 732.442727] env[62753]: INFO nova.compute.manager [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] [instance: efaf8c7c-3f38-4a18-97e8-099da390c7d8] Took 1.02 seconds to deallocate network for instance. [ 732.547633] env[62753]: DEBUG nova.compute.manager [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 732.732574] env[62753]: DEBUG nova.network.neutron [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 732.896726] env[62753]: DEBUG nova.network.neutron [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.071921] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 733.400116] env[62753]: DEBUG oslo_concurrency.lockutils [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "refresh_cache-3f46fa44-4410-4ca9-98d0-8efa4cf06d52" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 733.400633] env[62753]: DEBUG nova.compute.manager [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 733.400633] env[62753]: DEBUG nova.compute.manager [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 733.400828] env[62753]: DEBUG nova.network.neutron [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 733.419639] env[62753]: DEBUG nova.network.neutron [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 733.467871] env[62753]: INFO nova.scheduler.client.report [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Deleted allocations for instance efaf8c7c-3f38-4a18-97e8-099da390c7d8 [ 733.610962] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fddf9f41-c1e0-4699-ad01-944bf797ca68 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.619239] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-436d81f1-d7c5-44b6-96c9-39b4e89f917b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.654988] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8bfe437-14ba-4b9b-b324-2a80028cf363 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.663318] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37f1ac3c-4a00-4c97-9b94-097e589866a7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 733.676897] env[62753]: DEBUG nova.compute.provider_tree [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 733.923537] env[62753]: DEBUG nova.network.neutron [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 733.979164] env[62753]: DEBUG oslo_concurrency.lockutils [None req-399f0320-0722-43ff-abaa-27b5251ad19c tempest-ServersNegativeTestMultiTenantJSON-596689225 tempest-ServersNegativeTestMultiTenantJSON-596689225-project-member] Lock "efaf8c7c-3f38-4a18-97e8-099da390c7d8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.350s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.182859] env[62753]: DEBUG nova.scheduler.client.report [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 734.426762] env[62753]: INFO nova.compute.manager [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3f46fa44-4410-4ca9-98d0-8efa4cf06d52] Took 1.03 seconds to deallocate network for instance. [ 734.485518] env[62753]: DEBUG nova.compute.manager [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 734.687751] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 734.688773] env[62753]: DEBUG nova.compute.manager [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 734.691891] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.848s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 735.008912] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 735.199810] env[62753]: DEBUG nova.compute.utils [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 735.201382] env[62753]: DEBUG nova.compute.manager [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Not allocating networking since 'none' was specified. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 735.456923] env[62753]: INFO nova.scheduler.client.report [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Deleted allocations for instance 3f46fa44-4410-4ca9-98d0-8efa4cf06d52 [ 735.601087] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b10cbb07-dd4c-4f4a-98ca-c63c96bd8a35 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.609412] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bc61b63-8d41-498a-906d-0bf2a2f37cb4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.644728] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aaca2c1-9f07-45f2-8568-66d156c5041c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.653544] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b9689a1-a793-4c6c-bd0b-9a781ac5ca5a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 735.668591] env[62753]: DEBUG nova.compute.provider_tree [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 735.702809] env[62753]: DEBUG nova.compute.manager [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 735.966636] env[62753]: DEBUG oslo_concurrency.lockutils [None req-96a3d9d8-c181-4a81-b808-6dace1be566b tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "3f46fa44-4410-4ca9-98d0-8efa4cf06d52" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.927s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.172466] env[62753]: DEBUG nova.scheduler.client.report [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 736.470330] env[62753]: DEBUG nova.compute.manager [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 736.684161] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.992s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 736.684803] env[62753]: ERROR nova.compute.manager [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 78ea3c2f-4f6e-4bca-8b2b-d4a3ddd039ca, please check neutron logs for more information. [ 736.684803] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Traceback (most recent call last): [ 736.684803] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 736.684803] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] self.driver.spawn(context, instance, image_meta, [ 736.684803] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 736.684803] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] self._vmops.spawn(context, instance, image_meta, injected_files, [ 736.684803] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 736.684803] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] vm_ref = self.build_virtual_machine(instance, [ 736.684803] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 736.684803] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] vif_infos = vmwarevif.get_vif_info(self._session, [ 736.684803] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 736.685365] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] for vif in network_info: [ 736.685365] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 736.685365] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] return self._sync_wrapper(fn, *args, **kwargs) [ 736.685365] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 736.685365] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] self.wait() [ 736.685365] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 736.685365] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] self[:] = self._gt.wait() [ 736.685365] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 736.685365] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] return self._exit_event.wait() [ 736.685365] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 736.685365] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] result = hub.switch() [ 736.685365] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 736.685365] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] return self.greenlet.switch() [ 736.685947] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 736.685947] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] result = function(*args, **kwargs) [ 736.685947] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 736.685947] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] return func(*args, **kwargs) [ 736.685947] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 736.685947] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] raise e [ 736.685947] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 736.685947] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] nwinfo = self.network_api.allocate_for_instance( [ 736.685947] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 736.685947] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] created_port_ids = self._update_ports_for_instance( [ 736.685947] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 736.685947] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] with excutils.save_and_reraise_exception(): [ 736.685947] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 736.686340] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] self.force_reraise() [ 736.686340] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 736.686340] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] raise self.value [ 736.686340] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 736.686340] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] updated_port = self._update_port( [ 736.686340] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 736.686340] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] _ensure_no_port_binding_failure(port) [ 736.686340] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 736.686340] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] raise exception.PortBindingFailed(port_id=port['id']) [ 736.686340] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] nova.exception.PortBindingFailed: Binding failed for port 78ea3c2f-4f6e-4bca-8b2b-d4a3ddd039ca, please check neutron logs for more information. [ 736.686340] env[62753]: ERROR nova.compute.manager [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] [ 736.686673] env[62753]: DEBUG nova.compute.utils [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Binding failed for port 78ea3c2f-4f6e-4bca-8b2b-d4a3ddd039ca, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 736.687202] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 15.065s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 736.694342] env[62753]: DEBUG nova.compute.manager [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Build of instance 5ef5c03c-6c22-4db6-95ad-6c407c0bca78 was re-scheduled: Binding failed for port 78ea3c2f-4f6e-4bca-8b2b-d4a3ddd039ca, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 736.695493] env[62753]: DEBUG nova.compute.manager [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 736.695741] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Acquiring lock "refresh_cache-5ef5c03c-6c22-4db6-95ad-6c407c0bca78" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 736.695888] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Acquired lock "refresh_cache-5ef5c03c-6c22-4db6-95ad-6c407c0bca78" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 736.696416] env[62753]: DEBUG nova.network.neutron [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 736.715439] env[62753]: DEBUG nova.compute.manager [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 736.755741] env[62753]: DEBUG nova.virt.hardware [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 736.755974] env[62753]: DEBUG nova.virt.hardware [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 736.756191] env[62753]: DEBUG nova.virt.hardware [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 736.756418] env[62753]: DEBUG nova.virt.hardware [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 736.756607] env[62753]: DEBUG nova.virt.hardware [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 736.756785] env[62753]: DEBUG nova.virt.hardware [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 736.757150] env[62753]: DEBUG nova.virt.hardware [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 736.757331] env[62753]: DEBUG nova.virt.hardware [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 736.757504] env[62753]: DEBUG nova.virt.hardware [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 736.757667] env[62753]: DEBUG nova.virt.hardware [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 736.757837] env[62753]: DEBUG nova.virt.hardware [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 736.759715] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb60a66e-f464-4be1-a4d5-290a13fe7d51 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.768852] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c2a0785-13ce-4bf7-b857-726c33638043 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.783938] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Instance VIF info [] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 736.790614] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Creating folder: Project (121fec282e4c449ba5cea0101f504038). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 736.791084] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-cece4411-75c8-45de-8f0d-24f832017f60 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.803861] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Created folder: Project (121fec282e4c449ba5cea0101f504038) in parent group-v284541. [ 736.804227] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Creating folder: Instances. Parent ref: group-v284545. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 736.805362] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aa21456b-d90d-4b83-91e9-95a0274732e7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.815349] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Created folder: Instances in parent group-v284545. [ 736.816078] env[62753]: DEBUG oslo.service.loopingcall [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 736.816703] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 736.816961] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cbb7b6b5-1399-4e68-be31-49a27fbdb25d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 736.837134] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 736.837134] env[62753]: value = "task-1332089" [ 736.837134] env[62753]: _type = "Task" [ 736.837134] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 736.849046] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332089, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.007795] env[62753]: DEBUG oslo_concurrency.lockutils [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 737.308733] env[62753]: DEBUG nova.network.neutron [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 737.347300] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332089, 'name': CreateVM_Task, 'duration_secs': 0.270676} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 737.347644] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 737.348570] env[62753]: DEBUG oslo_vmware.service [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deabd9cc-4d8f-4e20-b2d6-6be6942182aa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.357075] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.357262] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.357622] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 737.358224] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-750e2bb0-917a-4678-8c05-aadc7b44bab3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.363247] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 737.363247] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]521e6b88-d068-855e-9414-ffe6f4b98ac5" [ 737.363247] env[62753]: _type = "Task" [ 737.363247] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.372158] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]521e6b88-d068-855e-9414-ffe6f4b98ac5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.491143] env[62753]: DEBUG nova.network.neutron [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 737.659221] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad25e439-5426-44e3-bf21-0a7e263289e4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.668141] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb3646e9-1efc-4fc5-a615-6f51bda8ebab {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.706680] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf21bbf-996c-4f49-b273-bc536c5b1a88 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.716504] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebba2b14-1f49-49f6-a07c-d5a2da6dcf79 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.733690] env[62753]: DEBUG nova.compute.provider_tree [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 737.877384] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.877645] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 737.877877] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 737.878141] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 737.878326] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 737.878586] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e1ea627-9934-48ce-aa2e-6367047f1924 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.899520] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 737.899520] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 737.899520] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83faf38e-6a25-4d00-bce6-b2f32423a592 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.908880] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-99dadf8d-6790-4f0f-8576-eb5ea118d356 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 737.914300] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 737.914300] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ec126f-5ff1-6caa-022f-bd28f528a156" [ 737.914300] env[62753]: _type = "Task" [ 737.914300] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 737.922935] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ec126f-5ff1-6caa-022f-bd28f528a156, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 737.993723] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Releasing lock "refresh_cache-5ef5c03c-6c22-4db6-95ad-6c407c0bca78" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 737.996020] env[62753]: DEBUG nova.compute.manager [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 737.996020] env[62753]: DEBUG nova.compute.manager [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 737.996020] env[62753]: DEBUG nova.network.neutron [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 738.019792] env[62753]: DEBUG nova.network.neutron [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 738.238987] env[62753]: DEBUG nova.scheduler.client.report [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 738.428165] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Preparing fetch location {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 738.431481] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Creating directory with path [datastore1] vmware_temp/eec6e6e1-9863-4ce1-ae0f-752e3ba8dbce/a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 738.431481] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-5e94f2f1-51f9-4db9-b0f2-0ac34e627f28 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.449136] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Created directory with path [datastore1] vmware_temp/eec6e6e1-9863-4ce1-ae0f-752e3ba8dbce/a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 738.449407] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Fetch image to [datastore1] vmware_temp/eec6e6e1-9863-4ce1-ae0f-752e3ba8dbce/a7541f15-bad6-4ea8-95ce-3499a4f01dda/tmp-sparse.vmdk {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 738.449600] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Downloading image file data a7541f15-bad6-4ea8-95ce-3499a4f01dda to [datastore1] vmware_temp/eec6e6e1-9863-4ce1-ae0f-752e3ba8dbce/a7541f15-bad6-4ea8-95ce-3499a4f01dda/tmp-sparse.vmdk on the data store datastore1 {{(pid=62753) _fetch_image_as_file /opt/stack/nova/nova/virt/vmwareapi/vmops.py:399}} [ 738.450431] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-52dac4cf-ed8a-42c4-96b2-ee6b76d94aea {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.458978] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c9ac254-5750-40f3-b4a2-8c5114edc311 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.472135] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1fd2bfc7-74e3-4fc9-b571-c289852971b6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.512642] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9105e116-312c-4414-9b5a-0743264a8dd2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.519530] env[62753]: DEBUG oslo_vmware.service [-] Invoking SessionManager.AcquireGenericServiceTicket with opID=oslo.vmware-f129f7e2-5a2c-4daf-b422-2349be2afe30 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 738.522891] env[62753]: DEBUG nova.network.neutron [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 738.550570] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Downloading image file data a7541f15-bad6-4ea8-95ce-3499a4f01dda to the data store datastore1 {{(pid=62753) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:245}} [ 738.616128] env[62753]: DEBUG oslo_vmware.rw_handles [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Creating HTTP connection to write to file with size = 21318656 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/eec6e6e1-9863-4ce1-ae0f-752e3ba8dbce/a7541f15-bad6-4ea8-95ce-3499a4f01dda/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62753) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 738.748076] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.061s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 738.748708] env[62753]: ERROR nova.compute.manager [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c2f33405-b9de-41af-9eed-37383de235f0, please check neutron logs for more information. [ 738.748708] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Traceback (most recent call last): [ 738.748708] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 738.748708] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] self.driver.spawn(context, instance, image_meta, [ 738.748708] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 738.748708] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] self._vmops.spawn(context, instance, image_meta, injected_files, [ 738.748708] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 738.748708] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] vm_ref = self.build_virtual_machine(instance, [ 738.748708] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 738.748708] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] vif_infos = vmwarevif.get_vif_info(self._session, [ 738.748708] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 738.749096] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] for vif in network_info: [ 738.749096] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 738.749096] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] return self._sync_wrapper(fn, *args, **kwargs) [ 738.749096] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 738.749096] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] self.wait() [ 738.749096] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 738.749096] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] self[:] = self._gt.wait() [ 738.749096] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 738.749096] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] return self._exit_event.wait() [ 738.749096] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 738.749096] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] result = hub.switch() [ 738.749096] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 738.749096] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] return self.greenlet.switch() [ 738.749458] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 738.749458] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] result = function(*args, **kwargs) [ 738.749458] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 738.749458] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] return func(*args, **kwargs) [ 738.749458] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 738.749458] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] raise e [ 738.749458] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 738.749458] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] nwinfo = self.network_api.allocate_for_instance( [ 738.749458] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 738.749458] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] created_port_ids = self._update_ports_for_instance( [ 738.749458] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 738.749458] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] with excutils.save_and_reraise_exception(): [ 738.749458] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 738.749881] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] self.force_reraise() [ 738.749881] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 738.749881] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] raise self.value [ 738.749881] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 738.749881] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] updated_port = self._update_port( [ 738.749881] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 738.749881] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] _ensure_no_port_binding_failure(port) [ 738.749881] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 738.749881] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] raise exception.PortBindingFailed(port_id=port['id']) [ 738.749881] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] nova.exception.PortBindingFailed: Binding failed for port c2f33405-b9de-41af-9eed-37383de235f0, please check neutron logs for more information. [ 738.749881] env[62753]: ERROR nova.compute.manager [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] [ 738.750314] env[62753]: DEBUG nova.compute.utils [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Binding failed for port c2f33405-b9de-41af-9eed-37383de235f0, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 738.751305] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.004s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 738.755401] env[62753]: INFO nova.compute.claims [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 738.758246] env[62753]: DEBUG nova.compute.manager [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Build of instance 50a480ed-f320-45d2-8ac9-a42d01614d11 was re-scheduled: Binding failed for port c2f33405-b9de-41af-9eed-37383de235f0, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 738.758934] env[62753]: DEBUG nova.compute.manager [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 738.759268] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Acquiring lock "refresh_cache-50a480ed-f320-45d2-8ac9-a42d01614d11" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 738.759431] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Acquired lock "refresh_cache-50a480ed-f320-45d2-8ac9-a42d01614d11" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 738.759627] env[62753]: DEBUG nova.network.neutron [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 738.948741] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "16c82d55-1420-4f83-a547-295e793de9df" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 738.948996] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "16c82d55-1420-4f83-a547-295e793de9df" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.025136] env[62753]: INFO nova.compute.manager [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] [instance: 5ef5c03c-6c22-4db6-95ad-6c407c0bca78] Took 1.03 seconds to deallocate network for instance. [ 739.149485] env[62753]: DEBUG oslo_concurrency.lockutils [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "3bc05fdb-81be-4764-b9d9-01acd125b020" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 739.149732] env[62753]: DEBUG oslo_concurrency.lockutils [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "3bc05fdb-81be-4764-b9d9-01acd125b020" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 739.284337] env[62753]: DEBUG nova.network.neutron [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.295674] env[62753]: DEBUG oslo_vmware.rw_handles [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Completed reading data from the image iterator. {{(pid=62753) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 739.295878] env[62753]: DEBUG oslo_vmware.rw_handles [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Closing write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap:443/folder/vmware_temp/eec6e6e1-9863-4ce1-ae0f-752e3ba8dbce/a7541f15-bad6-4ea8-95ce-3499a4f01dda/tmp-sparse.vmdk?dcPath=ha-datacenter&dsName=datastore1. {{(pid=62753) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:281}} [ 739.334530] env[62753]: DEBUG nova.network.neutron [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 739.430350] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Downloaded image file data a7541f15-bad6-4ea8-95ce-3499a4f01dda to vmware_temp/eec6e6e1-9863-4ce1-ae0f-752e3ba8dbce/a7541f15-bad6-4ea8-95ce-3499a4f01dda/tmp-sparse.vmdk on the data store datastore1 {{(pid=62753) fetch_image /opt/stack/nova/nova/virt/vmwareapi/images.py:258}} [ 739.432679] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Caching image {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 739.432941] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Copying Virtual Disk [datastore1] vmware_temp/eec6e6e1-9863-4ce1-ae0f-752e3ba8dbce/a7541f15-bad6-4ea8-95ce-3499a4f01dda/tmp-sparse.vmdk to [datastore1] vmware_temp/eec6e6e1-9863-4ce1-ae0f-752e3ba8dbce/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 739.433241] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-55f8eab6-c883-462f-9574-2cccff880329 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 739.441813] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 739.441813] env[62753]: value = "task-1332090" [ 739.441813] env[62753]: _type = "Task" [ 739.441813] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 739.449942] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332090, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 739.837920] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Releasing lock "refresh_cache-50a480ed-f320-45d2-8ac9-a42d01614d11" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 739.837920] env[62753]: DEBUG nova.compute.manager [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 739.837920] env[62753]: DEBUG nova.compute.manager [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 739.837920] env[62753]: DEBUG nova.network.neutron [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 739.868428] env[62753]: DEBUG nova.network.neutron [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 739.956270] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332090, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.058798] env[62753]: INFO nova.scheduler.client.report [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Deleted allocations for instance 5ef5c03c-6c22-4db6-95ad-6c407c0bca78 [ 740.181836] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95989daf-21f2-4b80-b7b9-20abb2f500ca {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.190356] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa5c2577-84a7-4d54-af60-86fc53de7126 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.221155] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af439820-20bc-4112-bf2c-88f5ffa4cc95 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.228462] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88c4ce74-0bf5-4440-9e05-9db6866792b0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.243573] env[62753]: DEBUG nova.compute.provider_tree [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 740.375573] env[62753]: DEBUG nova.network.neutron [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 740.453283] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332090, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.719941} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.453542] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Copied Virtual Disk [datastore1] vmware_temp/eec6e6e1-9863-4ce1-ae0f-752e3ba8dbce/a7541f15-bad6-4ea8-95ce-3499a4f01dda/tmp-sparse.vmdk to [datastore1] vmware_temp/eec6e6e1-9863-4ce1-ae0f-752e3ba8dbce/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 740.453732] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Deleting the datastore file [datastore1] vmware_temp/eec6e6e1-9863-4ce1-ae0f-752e3ba8dbce/a7541f15-bad6-4ea8-95ce-3499a4f01dda/tmp-sparse.vmdk {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 740.453969] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ed1e4f49-1802-4347-81ff-353a1b799694 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.459723] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 740.459723] env[62753]: value = "task-1332091" [ 740.459723] env[62753]: _type = "Task" [ 740.459723] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.467426] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332091, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 740.566893] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc78705f-13cc-4162-9f0f-732ef36b3eaf tempest-ServerPasswordTestJSON-472718264 tempest-ServerPasswordTestJSON-472718264-project-member] Lock "5ef5c03c-6c22-4db6-95ad-6c407c0bca78" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 116.234s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 740.748722] env[62753]: DEBUG nova.scheduler.client.report [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 740.877196] env[62753]: INFO nova.compute.manager [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] [instance: 50a480ed-f320-45d2-8ac9-a42d01614d11] Took 1.04 seconds to deallocate network for instance. [ 740.970144] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332091, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.021214} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 740.970387] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 740.970595] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Moving file from [datastore1] vmware_temp/eec6e6e1-9863-4ce1-ae0f-752e3ba8dbce/a7541f15-bad6-4ea8-95ce-3499a4f01dda to [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda. {{(pid=62753) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:334}} [ 740.970840] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MoveDatastoreFile_Task with opID=oslo.vmware-3935d1ae-0cd9-4117-b8f7-47d4492f4fc0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 740.978070] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 740.978070] env[62753]: value = "task-1332092" [ 740.978070] env[62753]: _type = "Task" [ 740.978070] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 740.985453] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332092, 'name': MoveDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.070575] env[62753]: DEBUG nova.compute.manager [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 741.255060] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.503s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 741.255580] env[62753]: DEBUG nova.compute.manager [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 741.258827] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.281s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 741.487381] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332092, 'name': MoveDatastoreFile_Task, 'duration_secs': 0.023814} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 741.487623] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] File moved {{(pid=62753) file_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:346}} [ 741.487814] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Cleaning up location [datastore1] vmware_temp/eec6e6e1-9863-4ce1-ae0f-752e3ba8dbce {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 741.487976] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Deleting the datastore file [datastore1] vmware_temp/eec6e6e1-9863-4ce1-ae0f-752e3ba8dbce {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 741.488233] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5447b2e6-6f03-4a40-bc9b-8c9f3abdaadf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 741.494200] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 741.494200] env[62753]: value = "task-1332093" [ 741.494200] env[62753]: _type = "Task" [ 741.494200] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 741.502564] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332093, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 741.591351] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 741.760475] env[62753]: DEBUG nova.compute.utils [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 741.761969] env[62753]: DEBUG nova.compute.manager [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 741.762619] env[62753]: DEBUG nova.network.neutron [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 741.907880] env[62753]: INFO nova.scheduler.client.report [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Deleted allocations for instance 50a480ed-f320-45d2-8ac9-a42d01614d11 [ 742.009468] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332093, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.022412} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.009750] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 742.010519] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39d57f95-0984-4e2d-9135-4721eaa515b1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.014429] env[62753]: DEBUG nova.policy [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '27a27d7d5034428887b4a487e81c3bf5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ee14ea30e3c94ddcaeed1ffbf32cde30', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 742.019596] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 742.019596] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]520fc1d7-5042-6da6-e924-ccb1c0bb2ec9" [ 742.019596] env[62753]: _type = "Task" [ 742.019596] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.031323] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]520fc1d7-5042-6da6-e924-ccb1c0bb2ec9, 'name': SearchDatastore_Task, 'duration_secs': 0.008644} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 742.031717] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 742.032089] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] a8ceac07-6631-4ac6-b62a-7f3eac788f2b/a8ceac07-6631-4ac6-b62a-7f3eac788f2b.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 742.032430] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a08499d3-f02a-4e81-8720-3518cfe7385c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.041861] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 742.041861] env[62753]: value = "task-1332094" [ 742.041861] env[62753]: _type = "Task" [ 742.041861] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 742.050161] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332094, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.146750] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f28178e6-092d-4223-95c4-1902e8a0513f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.154733] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7835a06-edbd-4589-8544-d0c596a2b20d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.189490] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a807231-c26b-4139-88d8-01294c0eb193 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.197502] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae1d6ec-ce3d-4895-afad-afe30437d3ef {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 742.215294] env[62753]: DEBUG nova.compute.provider_tree [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 742.265559] env[62753]: DEBUG nova.compute.manager [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 742.420105] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d1e04262-4868-49cd-93d2-3fd6ed9569cb tempest-ServerExternalEventsTest-1602293620 tempest-ServerExternalEventsTest-1602293620-project-member] Lock "50a480ed-f320-45d2-8ac9-a42d01614d11" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.985s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 742.559334] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332094, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 742.683197] env[62753]: DEBUG nova.network.neutron [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Successfully created port: bd51ad15-871e-4a77-bdea-d828df308468 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 742.718476] env[62753]: DEBUG nova.scheduler.client.report [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 742.923437] env[62753]: DEBUG nova.compute.manager [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 743.055217] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332094, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.224356] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.966s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 743.225072] env[62753]: ERROR nova.compute.manager [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port c8a4dec9-bd48-48eb-900d-3dee1f1aa326, please check neutron logs for more information. [ 743.225072] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Traceback (most recent call last): [ 743.225072] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 743.225072] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] self.driver.spawn(context, instance, image_meta, [ 743.225072] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 743.225072] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] self._vmops.spawn(context, instance, image_meta, injected_files, [ 743.225072] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 743.225072] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] vm_ref = self.build_virtual_machine(instance, [ 743.225072] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 743.225072] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] vif_infos = vmwarevif.get_vif_info(self._session, [ 743.225072] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 743.225588] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] for vif in network_info: [ 743.225588] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 743.225588] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] return self._sync_wrapper(fn, *args, **kwargs) [ 743.225588] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 743.225588] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] self.wait() [ 743.225588] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 743.225588] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] self[:] = self._gt.wait() [ 743.225588] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 743.225588] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] return self._exit_event.wait() [ 743.225588] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 743.225588] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] result = hub.switch() [ 743.225588] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 743.225588] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] return self.greenlet.switch() [ 743.225946] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 743.225946] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] result = function(*args, **kwargs) [ 743.225946] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 743.225946] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] return func(*args, **kwargs) [ 743.225946] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 743.225946] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] raise e [ 743.225946] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 743.225946] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] nwinfo = self.network_api.allocate_for_instance( [ 743.225946] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 743.225946] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] created_port_ids = self._update_ports_for_instance( [ 743.225946] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 743.225946] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] with excutils.save_and_reraise_exception(): [ 743.225946] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 743.226288] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] self.force_reraise() [ 743.226288] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 743.226288] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] raise self.value [ 743.226288] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 743.226288] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] updated_port = self._update_port( [ 743.226288] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 743.226288] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] _ensure_no_port_binding_failure(port) [ 743.226288] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 743.226288] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] raise exception.PortBindingFailed(port_id=port['id']) [ 743.226288] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] nova.exception.PortBindingFailed: Binding failed for port c8a4dec9-bd48-48eb-900d-3dee1f1aa326, please check neutron logs for more information. [ 743.226288] env[62753]: ERROR nova.compute.manager [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] [ 743.226552] env[62753]: DEBUG nova.compute.utils [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Binding failed for port c8a4dec9-bd48-48eb-900d-3dee1f1aa326, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 743.226977] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.572s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 743.231309] env[62753]: DEBUG nova.compute.manager [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Build of instance 250f2711-6f32-4c4d-9876-35a3a1c30d12 was re-scheduled: Binding failed for port c8a4dec9-bd48-48eb-900d-3dee1f1aa326, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 743.232351] env[62753]: DEBUG nova.compute.manager [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 743.232615] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Acquiring lock "refresh_cache-250f2711-6f32-4c4d-9876-35a3a1c30d12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.232760] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Acquired lock "refresh_cache-250f2711-6f32-4c4d-9876-35a3a1c30d12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.232992] env[62753]: DEBUG nova.network.neutron [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 743.276386] env[62753]: DEBUG nova.compute.manager [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 743.308246] env[62753]: DEBUG nova.virt.hardware [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 743.308785] env[62753]: DEBUG nova.virt.hardware [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 743.310048] env[62753]: DEBUG nova.virt.hardware [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 743.310048] env[62753]: DEBUG nova.virt.hardware [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 743.310048] env[62753]: DEBUG nova.virt.hardware [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 743.310048] env[62753]: DEBUG nova.virt.hardware [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 743.310048] env[62753]: DEBUG nova.virt.hardware [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 743.310281] env[62753]: DEBUG nova.virt.hardware [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 743.310281] env[62753]: DEBUG nova.virt.hardware [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 743.310281] env[62753]: DEBUG nova.virt.hardware [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 743.310392] env[62753]: DEBUG nova.virt.hardware [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 743.311247] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f6e14c3-c76c-4958-b0bb-1368f3a796dc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.321458] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b6c17dc-10df-4aac-8997-dcad87fcc862 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 743.452553] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 743.565681] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332094, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 743.764207] env[62753]: DEBUG nova.network.neutron [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 743.877476] env[62753]: DEBUG nova.network.neutron [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 743.907910] env[62753]: DEBUG nova.compute.manager [req-dbc8e91a-b7d1-4783-b1b8-c476929a9b85 req-0de994d6-f7b4-4cd1-9218-a4d870e14d79 service nova] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Received event network-changed-bd51ad15-871e-4a77-bdea-d828df308468 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 743.908139] env[62753]: DEBUG nova.compute.manager [req-dbc8e91a-b7d1-4783-b1b8-c476929a9b85 req-0de994d6-f7b4-4cd1-9218-a4d870e14d79 service nova] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Refreshing instance network info cache due to event network-changed-bd51ad15-871e-4a77-bdea-d828df308468. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 743.908372] env[62753]: DEBUG oslo_concurrency.lockutils [req-dbc8e91a-b7d1-4783-b1b8-c476929a9b85 req-0de994d6-f7b4-4cd1-9218-a4d870e14d79 service nova] Acquiring lock "refresh_cache-da9b45e9-618d-4e1e-9c49-3d702b2435e9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 743.908518] env[62753]: DEBUG oslo_concurrency.lockutils [req-dbc8e91a-b7d1-4783-b1b8-c476929a9b85 req-0de994d6-f7b4-4cd1-9218-a4d870e14d79 service nova] Acquired lock "refresh_cache-da9b45e9-618d-4e1e-9c49-3d702b2435e9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 743.908706] env[62753]: DEBUG nova.network.neutron [req-dbc8e91a-b7d1-4783-b1b8-c476929a9b85 req-0de994d6-f7b4-4cd1-9218-a4d870e14d79 service nova] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Refreshing network info cache for port bd51ad15-871e-4a77-bdea-d828df308468 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 744.062639] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332094, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.116114] env[62753]: ERROR nova.compute.manager [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port bd51ad15-871e-4a77-bdea-d828df308468, please check neutron logs for more information. [ 744.116114] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 744.116114] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.116114] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 744.116114] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.116114] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 744.116114] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.116114] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 744.116114] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.116114] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 744.116114] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.116114] env[62753]: ERROR nova.compute.manager raise self.value [ 744.116114] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.116114] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 744.116114] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.116114] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 744.116635] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.116635] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 744.116635] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port bd51ad15-871e-4a77-bdea-d828df308468, please check neutron logs for more information. [ 744.116635] env[62753]: ERROR nova.compute.manager [ 744.116635] env[62753]: Traceback (most recent call last): [ 744.116635] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 744.116635] env[62753]: listener.cb(fileno) [ 744.116635] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.116635] env[62753]: result = function(*args, **kwargs) [ 744.116635] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 744.116635] env[62753]: return func(*args, **kwargs) [ 744.116635] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 744.116635] env[62753]: raise e [ 744.116635] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.116635] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 744.116635] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.116635] env[62753]: created_port_ids = self._update_ports_for_instance( [ 744.116635] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.116635] env[62753]: with excutils.save_and_reraise_exception(): [ 744.116635] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.116635] env[62753]: self.force_reraise() [ 744.116635] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.116635] env[62753]: raise self.value [ 744.116635] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.116635] env[62753]: updated_port = self._update_port( [ 744.116635] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.116635] env[62753]: _ensure_no_port_binding_failure(port) [ 744.116635] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.116635] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 744.118399] env[62753]: nova.exception.PortBindingFailed: Binding failed for port bd51ad15-871e-4a77-bdea-d828df308468, please check neutron logs for more information. [ 744.118399] env[62753]: Removing descriptor: 14 [ 744.118399] env[62753]: ERROR nova.compute.manager [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port bd51ad15-871e-4a77-bdea-d828df308468, please check neutron logs for more information. [ 744.118399] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Traceback (most recent call last): [ 744.118399] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 744.118399] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] yield resources [ 744.118399] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 744.118399] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] self.driver.spawn(context, instance, image_meta, [ 744.118399] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 744.118399] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 744.118399] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 744.118399] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] vm_ref = self.build_virtual_machine(instance, [ 744.118742] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 744.118742] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] vif_infos = vmwarevif.get_vif_info(self._session, [ 744.118742] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 744.118742] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] for vif in network_info: [ 744.118742] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 744.118742] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] return self._sync_wrapper(fn, *args, **kwargs) [ 744.118742] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 744.118742] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] self.wait() [ 744.118742] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 744.118742] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] self[:] = self._gt.wait() [ 744.118742] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 744.118742] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] return self._exit_event.wait() [ 744.118742] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 744.119492] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] result = hub.switch() [ 744.119492] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 744.119492] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] return self.greenlet.switch() [ 744.119492] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 744.119492] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] result = function(*args, **kwargs) [ 744.119492] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 744.119492] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] return func(*args, **kwargs) [ 744.119492] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 744.119492] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] raise e [ 744.119492] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 744.119492] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] nwinfo = self.network_api.allocate_for_instance( [ 744.119492] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 744.119492] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] created_port_ids = self._update_ports_for_instance( [ 744.120077] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 744.120077] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] with excutils.save_and_reraise_exception(): [ 744.120077] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 744.120077] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] self.force_reraise() [ 744.120077] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 744.120077] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] raise self.value [ 744.120077] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 744.120077] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] updated_port = self._update_port( [ 744.120077] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 744.120077] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] _ensure_no_port_binding_failure(port) [ 744.120077] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 744.120077] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] raise exception.PortBindingFailed(port_id=port['id']) [ 744.120584] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] nova.exception.PortBindingFailed: Binding failed for port bd51ad15-871e-4a77-bdea-d828df308468, please check neutron logs for more information. [ 744.120584] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] [ 744.120584] env[62753]: INFO nova.compute.manager [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Terminating instance [ 744.121598] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Acquiring lock "refresh_cache-da9b45e9-618d-4e1e-9c49-3d702b2435e9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 744.172841] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43064fbe-4740-4c7f-ba9d-d766fea6743a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.180966] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-949425ec-0e50-4240-a916-2f0dbcffa973 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.213717] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13efabca-19f0-4819-9b71-55c493fa7ac9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.222685] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-625f95c2-4ce9-4c82-8196-6259aef422fe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.235995] env[62753]: DEBUG nova.compute.provider_tree [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 744.382722] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Releasing lock "refresh_cache-250f2711-6f32-4c4d-9876-35a3a1c30d12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 744.382979] env[62753]: DEBUG nova.compute.manager [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 744.383221] env[62753]: DEBUG nova.compute.manager [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 744.383392] env[62753]: DEBUG nova.network.neutron [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 744.408297] env[62753]: DEBUG nova.network.neutron [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.429961] env[62753]: DEBUG nova.network.neutron [req-dbc8e91a-b7d1-4783-b1b8-c476929a9b85 req-0de994d6-f7b4-4cd1-9218-a4d870e14d79 service nova] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 744.533692] env[62753]: DEBUG nova.network.neutron [req-dbc8e91a-b7d1-4783-b1b8-c476929a9b85 req-0de994d6-f7b4-4cd1-9218-a4d870e14d79 service nova] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 744.558618] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332094, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.279062} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 744.558878] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] a8ceac07-6631-4ac6-b62a-7f3eac788f2b/a8ceac07-6631-4ac6-b62a-7f3eac788f2b.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 744.559121] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 744.559388] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6abf11ae-eb46-43ee-8810-15e51d060c60 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 744.566121] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 744.566121] env[62753]: value = "task-1332095" [ 744.566121] env[62753]: _type = "Task" [ 744.566121] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 744.573787] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332095, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 744.738794] env[62753]: DEBUG nova.scheduler.client.report [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 744.911149] env[62753]: DEBUG nova.network.neutron [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.037638] env[62753]: DEBUG oslo_concurrency.lockutils [req-dbc8e91a-b7d1-4783-b1b8-c476929a9b85 req-0de994d6-f7b4-4cd1-9218-a4d870e14d79 service nova] Releasing lock "refresh_cache-da9b45e9-618d-4e1e-9c49-3d702b2435e9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 745.037638] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Acquired lock "refresh_cache-da9b45e9-618d-4e1e-9c49-3d702b2435e9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.037810] env[62753]: DEBUG nova.network.neutron [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 745.076042] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332095, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074084} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 745.076318] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 745.077098] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21f584bf-b198-49dd-a394-de7253307c2c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.096757] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Reconfiguring VM instance instance-00000023 to attach disk [datastore1] a8ceac07-6631-4ac6-b62a-7f3eac788f2b/a8ceac07-6631-4ac6-b62a-7f3eac788f2b.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 745.097047] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bda66c98-62c4-43ef-bd53-97caf82afe07 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 745.116390] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 745.116390] env[62753]: value = "task-1332096" [ 745.116390] env[62753]: _type = "Task" [ 745.116390] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 745.124809] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332096, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.247525] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 2.020s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 745.248207] env[62753]: ERROR nova.compute.manager [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port db911c7c-0d38-4274-b8c2-1c40aa502460, please check neutron logs for more information. [ 745.248207] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Traceback (most recent call last): [ 745.248207] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 745.248207] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] self.driver.spawn(context, instance, image_meta, [ 745.248207] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 745.248207] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 745.248207] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 745.248207] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] vm_ref = self.build_virtual_machine(instance, [ 745.248207] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 745.248207] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] vif_infos = vmwarevif.get_vif_info(self._session, [ 745.248207] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 745.248542] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] for vif in network_info: [ 745.248542] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 745.248542] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] return self._sync_wrapper(fn, *args, **kwargs) [ 745.248542] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 745.248542] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] self.wait() [ 745.248542] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 745.248542] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] self[:] = self._gt.wait() [ 745.248542] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 745.248542] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] return self._exit_event.wait() [ 745.248542] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 745.248542] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] result = hub.switch() [ 745.248542] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 745.248542] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] return self.greenlet.switch() [ 745.248853] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 745.248853] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] result = function(*args, **kwargs) [ 745.248853] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 745.248853] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] return func(*args, **kwargs) [ 745.248853] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 745.248853] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] raise e [ 745.248853] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 745.248853] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] nwinfo = self.network_api.allocate_for_instance( [ 745.248853] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 745.248853] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] created_port_ids = self._update_ports_for_instance( [ 745.248853] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 745.248853] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] with excutils.save_and_reraise_exception(): [ 745.248853] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 745.249183] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] self.force_reraise() [ 745.249183] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 745.249183] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] raise self.value [ 745.249183] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 745.249183] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] updated_port = self._update_port( [ 745.249183] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 745.249183] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] _ensure_no_port_binding_failure(port) [ 745.249183] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 745.249183] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] raise exception.PortBindingFailed(port_id=port['id']) [ 745.249183] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] nova.exception.PortBindingFailed: Binding failed for port db911c7c-0d38-4274-b8c2-1c40aa502460, please check neutron logs for more information. [ 745.249183] env[62753]: ERROR nova.compute.manager [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] [ 745.249468] env[62753]: DEBUG nova.compute.utils [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Binding failed for port db911c7c-0d38-4274-b8c2-1c40aa502460, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 745.250808] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.300s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 745.254115] env[62753]: DEBUG nova.compute.manager [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Build of instance a413174f-2a97-4c41-b512-e8ab562a25f4 was re-scheduled: Binding failed for port db911c7c-0d38-4274-b8c2-1c40aa502460, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 745.254652] env[62753]: DEBUG nova.compute.manager [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 745.254957] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Acquiring lock "refresh_cache-a413174f-2a97-4c41-b512-e8ab562a25f4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 745.255181] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Acquired lock "refresh_cache-a413174f-2a97-4c41-b512-e8ab562a25f4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 745.255398] env[62753]: DEBUG nova.network.neutron [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 745.414434] env[62753]: INFO nova.compute.manager [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] [instance: 250f2711-6f32-4c4d-9876-35a3a1c30d12] Took 1.03 seconds to deallocate network for instance. [ 745.555554] env[62753]: DEBUG nova.network.neutron [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.626029] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332096, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 745.638403] env[62753]: DEBUG nova.network.neutron [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.776182] env[62753]: DEBUG nova.network.neutron [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 745.872200] env[62753]: DEBUG nova.network.neutron [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 745.931531] env[62753]: DEBUG nova.compute.manager [req-6ef0a288-a123-40bf-99ac-c6f93a3aaf97 req-1c857acf-9fa2-4417-8413-67fa8f0582d1 service nova] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Received event network-vif-deleted-bd51ad15-871e-4a77-bdea-d828df308468 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 746.060527] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-177acc3f-4577-456a-b0a4-288490f32ff0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.068784] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09cd443b-a162-42a1-9248-dab97c68244a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.099541] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9ed349f-c82a-4648-a8e6-d70cd8720d06 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.107518] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8514a36d-ea9c-4356-a23c-41ff6407a3cc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.120210] env[62753]: DEBUG nova.compute.provider_tree [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 746.128769] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332096, 'name': ReconfigVM_Task, 'duration_secs': 0.764184} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.129591] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Reconfigured VM instance instance-00000023 to attach disk [datastore1] a8ceac07-6631-4ac6-b62a-7f3eac788f2b/a8ceac07-6631-4ac6-b62a-7f3eac788f2b.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 746.130179] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5efc4a51-a513-4f81-9b38-b0a14c9c1f1d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.136022] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 746.136022] env[62753]: value = "task-1332097" [ 746.136022] env[62753]: _type = "Task" [ 746.136022] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.140714] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Releasing lock "refresh_cache-da9b45e9-618d-4e1e-9c49-3d702b2435e9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.141108] env[62753]: DEBUG nova.compute.manager [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 746.141299] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 746.144107] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a450695c-91ac-457e-9a80-70a0b99e71a9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.145701] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332097, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.151312] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcbb33c2-a836-4737-b191-2386d8956bc7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.172112] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance da9b45e9-618d-4e1e-9c49-3d702b2435e9 could not be found. [ 746.172333] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 746.172514] env[62753]: INFO nova.compute.manager [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Took 0.03 seconds to destroy the instance on the hypervisor. [ 746.172756] env[62753]: DEBUG oslo.service.loopingcall [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 746.172969] env[62753]: DEBUG nova.compute.manager [-] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 746.173072] env[62753]: DEBUG nova.network.neutron [-] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 746.193349] env[62753]: DEBUG nova.network.neutron [-] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.375483] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Releasing lock "refresh_cache-a413174f-2a97-4c41-b512-e8ab562a25f4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 746.376458] env[62753]: DEBUG nova.compute.manager [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 746.376458] env[62753]: DEBUG nova.compute.manager [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 746.376458] env[62753]: DEBUG nova.network.neutron [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 746.396112] env[62753]: DEBUG nova.network.neutron [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 746.442226] env[62753]: INFO nova.scheduler.client.report [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Deleted allocations for instance 250f2711-6f32-4c4d-9876-35a3a1c30d12 [ 746.626322] env[62753]: DEBUG nova.scheduler.client.report [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 746.648074] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332097, 'name': Rename_Task, 'duration_secs': 0.138102} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 746.648344] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 746.649135] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8fe221eb-c289-44aa-a669-d3543e2aeb3d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 746.655897] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 746.655897] env[62753]: value = "task-1332098" [ 746.655897] env[62753]: _type = "Task" [ 746.655897] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 746.663149] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332098, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 746.696396] env[62753]: DEBUG nova.network.neutron [-] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.898862] env[62753]: DEBUG nova.network.neutron [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 746.952341] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24ccbea2-fa64-43a9-ad75-d131395b3b8b tempest-ServerMetadataTestJSON-1226411150 tempest-ServerMetadataTestJSON-1226411150-project-member] Lock "250f2711-6f32-4c4d-9876-35a3a1c30d12" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 117.135s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.132118] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.881s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 747.132502] env[62753]: ERROR nova.compute.manager [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 8b4fcaab-c1f4-4796-9d86-8de53459855a, please check neutron logs for more information. [ 747.132502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Traceback (most recent call last): [ 747.132502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 747.132502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] self.driver.spawn(context, instance, image_meta, [ 747.132502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 747.132502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] self._vmops.spawn(context, instance, image_meta, injected_files, [ 747.132502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 747.132502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] vm_ref = self.build_virtual_machine(instance, [ 747.132502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 747.132502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] vif_infos = vmwarevif.get_vif_info(self._session, [ 747.132502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 747.132852] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] for vif in network_info: [ 747.132852] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 747.132852] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] return self._sync_wrapper(fn, *args, **kwargs) [ 747.132852] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 747.132852] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] self.wait() [ 747.132852] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 747.132852] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] self[:] = self._gt.wait() [ 747.132852] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 747.132852] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] return self._exit_event.wait() [ 747.132852] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 747.132852] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] result = hub.switch() [ 747.132852] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 747.132852] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] return self.greenlet.switch() [ 747.133188] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 747.133188] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] result = function(*args, **kwargs) [ 747.133188] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 747.133188] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] return func(*args, **kwargs) [ 747.133188] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 747.133188] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] raise e [ 747.133188] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 747.133188] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] nwinfo = self.network_api.allocate_for_instance( [ 747.133188] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 747.133188] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] created_port_ids = self._update_ports_for_instance( [ 747.133188] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 747.133188] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] with excutils.save_and_reraise_exception(): [ 747.133188] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 747.133502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] self.force_reraise() [ 747.133502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 747.133502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] raise self.value [ 747.133502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 747.133502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] updated_port = self._update_port( [ 747.133502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 747.133502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] _ensure_no_port_binding_failure(port) [ 747.133502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 747.133502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] raise exception.PortBindingFailed(port_id=port['id']) [ 747.133502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] nova.exception.PortBindingFailed: Binding failed for port 8b4fcaab-c1f4-4796-9d86-8de53459855a, please check neutron logs for more information. [ 747.133502] env[62753]: ERROR nova.compute.manager [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] [ 747.133770] env[62753]: DEBUG nova.compute.utils [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Binding failed for port 8b4fcaab-c1f4-4796-9d86-8de53459855a, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 747.134453] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.063s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 747.135928] env[62753]: INFO nova.compute.claims [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 747.138677] env[62753]: DEBUG nova.compute.manager [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Build of instance c4ccc036-4963-4499-8bbb-dad5c0e4687e was re-scheduled: Binding failed for port 8b4fcaab-c1f4-4796-9d86-8de53459855a, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 747.138816] env[62753]: DEBUG nova.compute.manager [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 747.139036] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "refresh_cache-c4ccc036-4963-4499-8bbb-dad5c0e4687e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 747.139188] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired lock "refresh_cache-c4ccc036-4963-4499-8bbb-dad5c0e4687e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 747.139343] env[62753]: DEBUG nova.network.neutron [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 747.167479] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332098, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 747.199013] env[62753]: INFO nova.compute.manager [-] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Took 1.03 seconds to deallocate network for instance. [ 747.201297] env[62753]: DEBUG nova.compute.claims [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 747.201524] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 747.401389] env[62753]: INFO nova.compute.manager [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] [instance: a413174f-2a97-4c41-b512-e8ab562a25f4] Took 1.03 seconds to deallocate network for instance. [ 747.457404] env[62753]: DEBUG nova.compute.manager [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 747.662126] env[62753]: DEBUG nova.network.neutron [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 747.670312] env[62753]: DEBUG oslo_vmware.api [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332098, 'name': PowerOnVM_Task, 'duration_secs': 0.980413} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 747.670701] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 747.670995] env[62753]: INFO nova.compute.manager [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Took 10.96 seconds to spawn the instance on the hypervisor. [ 747.671469] env[62753]: DEBUG nova.compute.manager [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 747.672382] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1d652ac-4c84-441b-a10e-6fef8a50a72b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 747.808616] env[62753]: DEBUG nova.network.neutron [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 747.978441] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 748.190663] env[62753]: INFO nova.compute.manager [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Took 31.45 seconds to build instance. [ 748.313171] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Releasing lock "refresh_cache-c4ccc036-4963-4499-8bbb-dad5c0e4687e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 748.313481] env[62753]: DEBUG nova.compute.manager [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 748.313675] env[62753]: DEBUG nova.compute.manager [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 748.313859] env[62753]: DEBUG nova.network.neutron [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 748.333183] env[62753]: DEBUG nova.network.neutron [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 748.437383] env[62753]: INFO nova.scheduler.client.report [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Deleted allocations for instance a413174f-2a97-4c41-b512-e8ab562a25f4 [ 748.486933] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-033368af-70da-4b37-8159-a10a997689cf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.496606] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a30b6d5-11ab-4a3d-8a3f-df19e5f068f2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.530758] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-219cd7a1-6c8f-4dd2-b73b-fc2c100bd024 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.538559] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3839d81f-9dca-43a8-bccd-341a2b3cb033 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 748.553531] env[62753]: DEBUG nova.compute.provider_tree [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 748.695687] env[62753]: DEBUG oslo_concurrency.lockutils [None req-24376680-4d05-40a1-a33c-094a5ac273c7 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Lock "a8ceac07-6631-4ac6-b62a-7f3eac788f2b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 113.133s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 748.836270] env[62753]: DEBUG nova.network.neutron [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 748.951528] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcf75422-e30b-4b39-a2fa-221003ecc665 tempest-ServersTestFqdnHostnames-1603543605 tempest-ServersTestFqdnHostnames-1603543605-project-member] Lock "a413174f-2a97-4c41-b512-e8ab562a25f4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 118.460s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.057380] env[62753]: DEBUG nova.scheduler.client.report [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 749.202513] env[62753]: DEBUG nova.compute.manager [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 749.340797] env[62753]: INFO nova.compute.manager [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: c4ccc036-4963-4499-8bbb-dad5c0e4687e] Took 1.03 seconds to deallocate network for instance. [ 749.400465] env[62753]: INFO nova.compute.manager [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Rebuilding instance [ 749.442029] env[62753]: DEBUG nova.compute.manager [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 749.442989] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d932e30c-0437-4942-a045-e2a1fa332611 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.452384] env[62753]: DEBUG nova.compute.manager [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 749.563924] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.429s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 749.564489] env[62753]: DEBUG nova.compute.manager [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 749.567569] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 14.559s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 749.568989] env[62753]: INFO nova.compute.claims [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 749.729266] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 749.956067] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 749.956067] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6a520e3-ee59-4352-a36c-76739f92c9e6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 749.966920] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 749.966920] env[62753]: value = "task-1332099" [ 749.966920] env[62753]: _type = "Task" [ 749.966920] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 749.974287] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332099, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 749.978053] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 750.071018] env[62753]: DEBUG nova.compute.utils [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 750.071250] env[62753]: DEBUG nova.compute.manager [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 750.073144] env[62753]: DEBUG nova.network.neutron [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 750.134261] env[62753]: DEBUG nova.policy [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '352626dae8044175bb867e28793f5bdc', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '0fa84ce7c4c448ca85675f6ec85337ae', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 750.377767] env[62753]: INFO nova.scheduler.client.report [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Deleted allocations for instance c4ccc036-4963-4499-8bbb-dad5c0e4687e [ 750.477491] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332099, 'name': PowerOffVM_Task, 'duration_secs': 0.118924} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 750.477542] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 750.478089] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 750.479410] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0bd97c6-56e2-403d-885b-ff3aac718dc3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.486959] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 750.487228] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6b1b753a-2832-49a9-a007-c02022318de2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.511702] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 750.511928] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 750.512124] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Deleting the datastore file [datastore1] a8ceac07-6631-4ac6-b62a-7f3eac788f2b {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 750.512393] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-90e19e3d-d463-4eb8-a877-251826c33078 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.520029] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 750.520029] env[62753]: value = "task-1332101" [ 750.520029] env[62753]: _type = "Task" [ 750.520029] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 750.534021] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332101, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 750.577157] env[62753]: DEBUG nova.compute.manager [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 750.583023] env[62753]: DEBUG nova.network.neutron [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Successfully created port: 09388d08-50e1-48bc-a122-f3442c3ec2ca {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 750.885763] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b718c363-3caa-4d06-ae11-d530cd1d8505 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "c4ccc036-4963-4499-8bbb-dad5c0e4687e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 115.971s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 750.962514] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcac43a5-f17e-4a2d-969d-fca7365d6958 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 750.970854] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e893651-f71b-4ade-8e10-534f608c74d2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.003369] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfa586a8-f412-499f-9856-f17b3e6bd840 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.011066] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-467cd334-1ed8-4ae8-8fa6-f26585c44e87 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.024431] env[62753]: DEBUG nova.compute.provider_tree [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 751.033362] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332101, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.087905} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 751.033603] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 751.033776] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 751.033937] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 751.389613] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 751.530907] env[62753]: DEBUG nova.scheduler.client.report [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 751.537215] env[62753]: DEBUG nova.compute.manager [req-875683ca-d87d-4771-8535-a03c4586f585 req-0b9aebc6-aaa4-499b-ae7d-d82494b6cd11 service nova] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Received event network-changed-09388d08-50e1-48bc-a122-f3442c3ec2ca {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 751.537559] env[62753]: DEBUG nova.compute.manager [req-875683ca-d87d-4771-8535-a03c4586f585 req-0b9aebc6-aaa4-499b-ae7d-d82494b6cd11 service nova] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Refreshing instance network info cache due to event network-changed-09388d08-50e1-48bc-a122-f3442c3ec2ca. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 751.537910] env[62753]: DEBUG oslo_concurrency.lockutils [req-875683ca-d87d-4771-8535-a03c4586f585 req-0b9aebc6-aaa4-499b-ae7d-d82494b6cd11 service nova] Acquiring lock "refresh_cache-e755857a-e0bc-40bf-bcbb-a416d243442b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.538210] env[62753]: DEBUG oslo_concurrency.lockutils [req-875683ca-d87d-4771-8535-a03c4586f585 req-0b9aebc6-aaa4-499b-ae7d-d82494b6cd11 service nova] Acquired lock "refresh_cache-e755857a-e0bc-40bf-bcbb-a416d243442b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 751.538601] env[62753]: DEBUG nova.network.neutron [req-875683ca-d87d-4771-8535-a03c4586f585 req-0b9aebc6-aaa4-499b-ae7d-d82494b6cd11 service nova] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Refreshing network info cache for port 09388d08-50e1-48bc-a122-f3442c3ec2ca {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 751.594999] env[62753]: DEBUG nova.compute.manager [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 751.620410] env[62753]: DEBUG nova.virt.hardware [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 751.620661] env[62753]: DEBUG nova.virt.hardware [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 751.620816] env[62753]: DEBUG nova.virt.hardware [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 751.620998] env[62753]: DEBUG nova.virt.hardware [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 751.621452] env[62753]: DEBUG nova.virt.hardware [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 751.621707] env[62753]: DEBUG nova.virt.hardware [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 751.621937] env[62753]: DEBUG nova.virt.hardware [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 751.622121] env[62753]: DEBUG nova.virt.hardware [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 751.622296] env[62753]: DEBUG nova.virt.hardware [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 751.622459] env[62753]: DEBUG nova.virt.hardware [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 751.622751] env[62753]: DEBUG nova.virt.hardware [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 751.623670] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c012fe7-e36e-4767-94a3-f0581daf5142 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.635483] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92044c0d-8c70-4449-bc20-51b2662bfd5e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 751.795146] env[62753]: ERROR nova.compute.manager [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 09388d08-50e1-48bc-a122-f3442c3ec2ca, please check neutron logs for more information. [ 751.795146] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 751.795146] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 751.795146] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 751.795146] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.795146] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 751.795146] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.795146] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 751.795146] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.795146] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 751.795146] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.795146] env[62753]: ERROR nova.compute.manager raise self.value [ 751.795146] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.795146] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 751.795146] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.795146] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 751.795599] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.795599] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 751.795599] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 09388d08-50e1-48bc-a122-f3442c3ec2ca, please check neutron logs for more information. [ 751.795599] env[62753]: ERROR nova.compute.manager [ 751.795599] env[62753]: Traceback (most recent call last): [ 751.795599] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 751.795599] env[62753]: listener.cb(fileno) [ 751.795599] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.795599] env[62753]: result = function(*args, **kwargs) [ 751.795599] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 751.795599] env[62753]: return func(*args, **kwargs) [ 751.795599] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 751.795599] env[62753]: raise e [ 751.795599] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 751.795599] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 751.795599] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.795599] env[62753]: created_port_ids = self._update_ports_for_instance( [ 751.795599] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.795599] env[62753]: with excutils.save_and_reraise_exception(): [ 751.795599] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.795599] env[62753]: self.force_reraise() [ 751.795599] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.795599] env[62753]: raise self.value [ 751.795599] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.795599] env[62753]: updated_port = self._update_port( [ 751.795599] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.795599] env[62753]: _ensure_no_port_binding_failure(port) [ 751.795599] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.795599] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 751.796366] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 09388d08-50e1-48bc-a122-f3442c3ec2ca, please check neutron logs for more information. [ 751.796366] env[62753]: Removing descriptor: 14 [ 751.796366] env[62753]: ERROR nova.compute.manager [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 09388d08-50e1-48bc-a122-f3442c3ec2ca, please check neutron logs for more information. [ 751.796366] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Traceback (most recent call last): [ 751.796366] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 751.796366] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] yield resources [ 751.796366] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 751.796366] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] self.driver.spawn(context, instance, image_meta, [ 751.796366] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 751.796366] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 751.796366] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 751.796366] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] vm_ref = self.build_virtual_machine(instance, [ 751.796668] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 751.796668] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] vif_infos = vmwarevif.get_vif_info(self._session, [ 751.796668] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 751.796668] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] for vif in network_info: [ 751.796668] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 751.796668] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] return self._sync_wrapper(fn, *args, **kwargs) [ 751.796668] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 751.796668] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] self.wait() [ 751.796668] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 751.796668] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] self[:] = self._gt.wait() [ 751.796668] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 751.796668] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] return self._exit_event.wait() [ 751.796668] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 751.796994] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] result = hub.switch() [ 751.796994] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 751.796994] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] return self.greenlet.switch() [ 751.796994] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 751.796994] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] result = function(*args, **kwargs) [ 751.796994] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 751.796994] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] return func(*args, **kwargs) [ 751.796994] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 751.796994] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] raise e [ 751.796994] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 751.796994] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] nwinfo = self.network_api.allocate_for_instance( [ 751.796994] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 751.796994] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] created_port_ids = self._update_ports_for_instance( [ 751.797326] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 751.797326] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] with excutils.save_and_reraise_exception(): [ 751.797326] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 751.797326] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] self.force_reraise() [ 751.797326] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 751.797326] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] raise self.value [ 751.797326] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 751.797326] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] updated_port = self._update_port( [ 751.797326] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 751.797326] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] _ensure_no_port_binding_failure(port) [ 751.797326] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 751.797326] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] raise exception.PortBindingFailed(port_id=port['id']) [ 751.797633] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] nova.exception.PortBindingFailed: Binding failed for port 09388d08-50e1-48bc-a122-f3442c3ec2ca, please check neutron logs for more information. [ 751.797633] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] [ 751.797633] env[62753]: INFO nova.compute.manager [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Terminating instance [ 751.798457] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Acquiring lock "refresh_cache-e755857a-e0bc-40bf-bcbb-a416d243442b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 751.918627] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.041750] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.474s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 752.042363] env[62753]: DEBUG nova.compute.manager [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 752.046380] env[62753]: DEBUG oslo_concurrency.lockutils [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.039s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 752.048767] env[62753]: INFO nova.compute.claims [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 752.074640] env[62753]: DEBUG nova.network.neutron [req-875683ca-d87d-4771-8535-a03c4586f585 req-0b9aebc6-aaa4-499b-ae7d-d82494b6cd11 service nova] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 752.078421] env[62753]: DEBUG nova.virt.hardware [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 752.078862] env[62753]: DEBUG nova.virt.hardware [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 752.078862] env[62753]: DEBUG nova.virt.hardware [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 752.078961] env[62753]: DEBUG nova.virt.hardware [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 752.079100] env[62753]: DEBUG nova.virt.hardware [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 752.079252] env[62753]: DEBUG nova.virt.hardware [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 752.079480] env[62753]: DEBUG nova.virt.hardware [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 752.079655] env[62753]: DEBUG nova.virt.hardware [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 752.079826] env[62753]: DEBUG nova.virt.hardware [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 752.079989] env[62753]: DEBUG nova.virt.hardware [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 752.080188] env[62753]: DEBUG nova.virt.hardware [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 752.081297] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff816036-6a4a-4c39-9ccd-cad7cc4611f0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.090524] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d89fe3e-9f00-4792-b613-cc621a4d346f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.104052] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Instance VIF info [] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 752.109534] env[62753]: DEBUG oslo.service.loopingcall [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 752.109779] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 752.109982] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0efe16ee-2a97-4eff-8aa6-672a6bcda59e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.128122] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 752.128122] env[62753]: value = "task-1332102" [ 752.128122] env[62753]: _type = "Task" [ 752.128122] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.138408] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332102, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.202073] env[62753]: DEBUG nova.network.neutron [req-875683ca-d87d-4771-8535-a03c4586f585 req-0b9aebc6-aaa4-499b-ae7d-d82494b6cd11 service nova] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 752.548076] env[62753]: DEBUG nova.compute.utils [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 752.549502] env[62753]: DEBUG nova.compute.manager [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 752.549821] env[62753]: DEBUG nova.network.neutron [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 752.600492] env[62753]: DEBUG nova.policy [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3bc95026e38a4af9bb20133b2a37db75', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7ad54077bd3b4a3791de66dddc40bb5a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 752.638105] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332102, 'name': CreateVM_Task, 'duration_secs': 0.501709} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 752.638261] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 752.638752] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 752.638954] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.639330] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 752.639626] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74b27a39-88be-45b0-953f-a6499787a050 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 752.644100] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 752.644100] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525b210c-d3a2-03fc-68f3-14537f31f21c" [ 752.644100] env[62753]: _type = "Task" [ 752.644100] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 752.651965] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525b210c-d3a2-03fc-68f3-14537f31f21c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 752.705568] env[62753]: DEBUG oslo_concurrency.lockutils [req-875683ca-d87d-4771-8535-a03c4586f585 req-0b9aebc6-aaa4-499b-ae7d-d82494b6cd11 service nova] Releasing lock "refresh_cache-e755857a-e0bc-40bf-bcbb-a416d243442b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 752.706638] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Acquired lock "refresh_cache-e755857a-e0bc-40bf-bcbb-a416d243442b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 752.706638] env[62753]: DEBUG nova.network.neutron [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 752.933600] env[62753]: DEBUG nova.network.neutron [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Successfully created port: b067d0d2-dcbe-4852-920e-c9130d215bdc {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 752.975984] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "046fc3dc-a187-4fda-bc66-345e1226f83d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 752.976723] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "046fc3dc-a187-4fda-bc66-345e1226f83d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 753.056017] env[62753]: DEBUG nova.compute.manager [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 753.155898] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525b210c-d3a2-03fc-68f3-14537f31f21c, 'name': SearchDatastore_Task, 'duration_secs': 0.009792} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.156271] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.156556] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 753.159146] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 753.159146] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 753.159146] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 753.159801] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-20f710e5-2c28-4763-91dc-6dfdddf6784b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.167950] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 753.168312] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 753.169044] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7267a61-33b8-4929-9616-87220b16b370 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.175785] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 753.175785] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bd35a9-313b-6dbe-980e-2e7b1bd53380" [ 753.175785] env[62753]: _type = "Task" [ 753.175785] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.183206] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bd35a9-313b-6dbe-980e-2e7b1bd53380, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.228294] env[62753]: DEBUG nova.network.neutron [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.319574] env[62753]: DEBUG nova.network.neutron [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 753.429259] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6c485ec-5937-4acf-b9f7-09593c472736 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.436230] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3d05734-5f06-43d3-be43-641f9a2db416 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.473228] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9bea0aaa-4937-4ac5-a068-ec8d5c6c4651 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.480432] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fd986f0-4dd1-497c-8e93-7cfe624df61f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.494102] env[62753]: DEBUG nova.compute.provider_tree [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 753.553175] env[62753]: DEBUG nova.compute.manager [req-27f5b509-5556-4cf7-a1f8-1a63815d2f1f req-e894e854-b0e1-4abe-985e-39d111325a6a service nova] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Received event network-vif-deleted-09388d08-50e1-48bc-a122-f3442c3ec2ca {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 753.690020] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bd35a9-313b-6dbe-980e-2e7b1bd53380, 'name': SearchDatastore_Task, 'duration_secs': 0.007666} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 753.690020] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5e68dc3f-486c-4f24-a2d4-88c4364fa494 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.693194] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 753.693194] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e59b61-a2c5-f922-4a61-a2b24870be28" [ 753.693194] env[62753]: _type = "Task" [ 753.693194] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 753.700948] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e59b61-a2c5-f922-4a61-a2b24870be28, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 753.822898] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Releasing lock "refresh_cache-e755857a-e0bc-40bf-bcbb-a416d243442b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 753.823325] env[62753]: DEBUG nova.compute.manager [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 753.823522] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 753.824570] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-37be0cbd-b946-466a-8933-3bcd7c70458a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.833234] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f77ef992-76ae-4d65-9fd9-65e78b89c018 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 753.854709] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e755857a-e0bc-40bf-bcbb-a416d243442b could not be found. [ 753.854937] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 753.855139] env[62753]: INFO nova.compute.manager [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Took 0.03 seconds to destroy the instance on the hypervisor. [ 753.857048] env[62753]: DEBUG oslo.service.loopingcall [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 753.857048] env[62753]: DEBUG nova.compute.manager [-] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 753.857048] env[62753]: DEBUG nova.network.neutron [-] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 753.872825] env[62753]: DEBUG nova.network.neutron [-] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 753.997971] env[62753]: DEBUG nova.scheduler.client.report [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 754.065338] env[62753]: DEBUG nova.compute.manager [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 754.091305] env[62753]: DEBUG nova.virt.hardware [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 754.091690] env[62753]: DEBUG nova.virt.hardware [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 754.091911] env[62753]: DEBUG nova.virt.hardware [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 754.092162] env[62753]: DEBUG nova.virt.hardware [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 754.092367] env[62753]: DEBUG nova.virt.hardware [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 754.092566] env[62753]: DEBUG nova.virt.hardware [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 754.092884] env[62753]: DEBUG nova.virt.hardware [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 754.093124] env[62753]: DEBUG nova.virt.hardware [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 754.093359] env[62753]: DEBUG nova.virt.hardware [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 754.093584] env[62753]: DEBUG nova.virt.hardware [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 754.093817] env[62753]: DEBUG nova.virt.hardware [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 754.094749] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee0ea3ca-9a6c-4731-adf7-77cafbc4cb8b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.103308] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec773055-6c05-4f02-84fb-3d841ee6932c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.203539] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e59b61-a2c5-f922-4a61-a2b24870be28, 'name': SearchDatastore_Task, 'duration_secs': 0.008068} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.203880] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 754.204153] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] a8ceac07-6631-4ac6-b62a-7f3eac788f2b/a8ceac07-6631-4ac6-b62a-7f3eac788f2b.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 754.204414] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ef98c317-22a4-4785-a204-9f1826075cf5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.210772] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 754.210772] env[62753]: value = "task-1332103" [ 754.210772] env[62753]: _type = "Task" [ 754.210772] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.218496] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332103, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.326569] env[62753]: ERROR nova.compute.manager [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b067d0d2-dcbe-4852-920e-c9130d215bdc, please check neutron logs for more information. [ 754.326569] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 754.326569] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.326569] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 754.326569] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 754.326569] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 754.326569] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 754.326569] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 754.326569] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.326569] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 754.326569] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.326569] env[62753]: ERROR nova.compute.manager raise self.value [ 754.326569] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 754.326569] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 754.326569] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.326569] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 754.327040] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.327040] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 754.327040] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b067d0d2-dcbe-4852-920e-c9130d215bdc, please check neutron logs for more information. [ 754.327040] env[62753]: ERROR nova.compute.manager [ 754.327040] env[62753]: Traceback (most recent call last): [ 754.327040] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 754.327040] env[62753]: listener.cb(fileno) [ 754.327040] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 754.327040] env[62753]: result = function(*args, **kwargs) [ 754.327040] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 754.327040] env[62753]: return func(*args, **kwargs) [ 754.327040] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 754.327040] env[62753]: raise e [ 754.327040] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.327040] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 754.327040] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 754.327040] env[62753]: created_port_ids = self._update_ports_for_instance( [ 754.327040] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 754.327040] env[62753]: with excutils.save_and_reraise_exception(): [ 754.327040] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.327040] env[62753]: self.force_reraise() [ 754.327040] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.327040] env[62753]: raise self.value [ 754.327040] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 754.327040] env[62753]: updated_port = self._update_port( [ 754.327040] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.327040] env[62753]: _ensure_no_port_binding_failure(port) [ 754.327040] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.327040] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 754.327874] env[62753]: nova.exception.PortBindingFailed: Binding failed for port b067d0d2-dcbe-4852-920e-c9130d215bdc, please check neutron logs for more information. [ 754.327874] env[62753]: Removing descriptor: 14 [ 754.327874] env[62753]: ERROR nova.compute.manager [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b067d0d2-dcbe-4852-920e-c9130d215bdc, please check neutron logs for more information. [ 754.327874] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Traceback (most recent call last): [ 754.327874] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 754.327874] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] yield resources [ 754.327874] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 754.327874] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] self.driver.spawn(context, instance, image_meta, [ 754.327874] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 754.327874] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 754.327874] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 754.327874] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] vm_ref = self.build_virtual_machine(instance, [ 754.328295] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 754.328295] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] vif_infos = vmwarevif.get_vif_info(self._session, [ 754.328295] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 754.328295] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] for vif in network_info: [ 754.328295] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 754.328295] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] return self._sync_wrapper(fn, *args, **kwargs) [ 754.328295] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 754.328295] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] self.wait() [ 754.328295] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 754.328295] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] self[:] = self._gt.wait() [ 754.328295] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 754.328295] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] return self._exit_event.wait() [ 754.328295] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 754.328724] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] result = hub.switch() [ 754.328724] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 754.328724] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] return self.greenlet.switch() [ 754.328724] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 754.328724] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] result = function(*args, **kwargs) [ 754.328724] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 754.328724] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] return func(*args, **kwargs) [ 754.328724] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 754.328724] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] raise e [ 754.328724] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 754.328724] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] nwinfo = self.network_api.allocate_for_instance( [ 754.328724] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 754.328724] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] created_port_ids = self._update_ports_for_instance( [ 754.329094] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 754.329094] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] with excutils.save_and_reraise_exception(): [ 754.329094] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 754.329094] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] self.force_reraise() [ 754.329094] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 754.329094] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] raise self.value [ 754.329094] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 754.329094] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] updated_port = self._update_port( [ 754.329094] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 754.329094] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] _ensure_no_port_binding_failure(port) [ 754.329094] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 754.329094] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] raise exception.PortBindingFailed(port_id=port['id']) [ 754.329431] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] nova.exception.PortBindingFailed: Binding failed for port b067d0d2-dcbe-4852-920e-c9130d215bdc, please check neutron logs for more information. [ 754.329431] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] [ 754.329431] env[62753]: INFO nova.compute.manager [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Terminating instance [ 754.330571] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Acquiring lock "refresh_cache-eca070f0-dda0-4c51-b2d3-1251c517804a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 754.330837] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Acquired lock "refresh_cache-eca070f0-dda0-4c51-b2d3-1251c517804a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 754.331139] env[62753]: DEBUG nova.network.neutron [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 754.375328] env[62753]: DEBUG nova.network.neutron [-] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 754.503183] env[62753]: DEBUG oslo_concurrency.lockutils [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.456s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 754.507516] env[62753]: DEBUG nova.compute.manager [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 754.508459] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 12.916s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 754.511072] env[62753]: INFO nova.compute.claims [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 754.720836] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332103, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.452715} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 754.721126] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] a8ceac07-6631-4ac6-b62a-7f3eac788f2b/a8ceac07-6631-4ac6-b62a-7f3eac788f2b.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 754.721319] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 754.721604] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fa926c2f-58c9-4828-8199-0076ce7ff14f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 754.727937] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 754.727937] env[62753]: value = "task-1332104" [ 754.727937] env[62753]: _type = "Task" [ 754.727937] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 754.735011] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332104, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 754.847651] env[62753]: DEBUG nova.network.neutron [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 754.877910] env[62753]: INFO nova.compute.manager [-] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Took 1.02 seconds to deallocate network for instance. [ 754.880209] env[62753]: DEBUG nova.compute.claims [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 754.880380] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 754.950486] env[62753]: DEBUG nova.network.neutron [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 755.015894] env[62753]: DEBUG nova.compute.utils [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 755.019143] env[62753]: DEBUG nova.compute.manager [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 755.019314] env[62753]: DEBUG nova.network.neutron [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 755.062249] env[62753]: DEBUG nova.policy [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fac3f48c3afb4e11bcbdc93b5c6b64ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5bca83da77f348e28f34e3f25e4ddd80', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 755.239312] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332104, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06462} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.239931] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 755.240807] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fedb4b63-96b6-4e79-ac3d-6d9e0ee09248 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.262022] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Reconfiguring VM instance instance-00000023 to attach disk [datastore2] a8ceac07-6631-4ac6-b62a-7f3eac788f2b/a8ceac07-6631-4ac6-b62a-7f3eac788f2b.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 755.262022] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-635c8cc0-45cf-45b6-896e-b0c8fce47370 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.280534] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 755.280534] env[62753]: value = "task-1332105" [ 755.280534] env[62753]: _type = "Task" [ 755.280534] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.288852] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332105, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.396249] env[62753]: DEBUG nova.network.neutron [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Successfully created port: f53e6ca3-7555-4166-b8d6-3613594ac4f8 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 755.453994] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Releasing lock "refresh_cache-eca070f0-dda0-4c51-b2d3-1251c517804a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 755.454557] env[62753]: DEBUG nova.compute.manager [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 755.454794] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 755.455149] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2f0a6b50-b3ab-43a1-b60c-91bdfbe30a8b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.466519] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb6c0c0-b97a-42f8-9160-1d71b809dca5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.489767] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance eca070f0-dda0-4c51-b2d3-1251c517804a could not be found. [ 755.490051] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 755.490283] env[62753]: INFO nova.compute.manager [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Took 0.04 seconds to destroy the instance on the hypervisor. [ 755.490587] env[62753]: DEBUG oslo.service.loopingcall [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 755.490829] env[62753]: DEBUG nova.compute.manager [-] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 755.490951] env[62753]: DEBUG nova.network.neutron [-] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 755.509276] env[62753]: DEBUG nova.network.neutron [-] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 755.519534] env[62753]: DEBUG nova.compute.manager [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 755.595366] env[62753]: DEBUG nova.compute.manager [req-7593527c-637c-464b-8754-9c99ce1f165f req-5f0d1294-8f96-446a-b274-64e8ae67535a service nova] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Received event network-changed-b067d0d2-dcbe-4852-920e-c9130d215bdc {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 755.595366] env[62753]: DEBUG nova.compute.manager [req-7593527c-637c-464b-8754-9c99ce1f165f req-5f0d1294-8f96-446a-b274-64e8ae67535a service nova] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Refreshing instance network info cache due to event network-changed-b067d0d2-dcbe-4852-920e-c9130d215bdc. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 755.595366] env[62753]: DEBUG oslo_concurrency.lockutils [req-7593527c-637c-464b-8754-9c99ce1f165f req-5f0d1294-8f96-446a-b274-64e8ae67535a service nova] Acquiring lock "refresh_cache-eca070f0-dda0-4c51-b2d3-1251c517804a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 755.595366] env[62753]: DEBUG oslo_concurrency.lockutils [req-7593527c-637c-464b-8754-9c99ce1f165f req-5f0d1294-8f96-446a-b274-64e8ae67535a service nova] Acquired lock "refresh_cache-eca070f0-dda0-4c51-b2d3-1251c517804a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 755.595366] env[62753]: DEBUG nova.network.neutron [req-7593527c-637c-464b-8754-9c99ce1f165f req-5f0d1294-8f96-446a-b274-64e8ae67535a service nova] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Refreshing network info cache for port b067d0d2-dcbe-4852-920e-c9130d215bdc {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 755.792942] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332105, 'name': ReconfigVM_Task, 'duration_secs': 0.283892} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 755.796070] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Reconfigured VM instance instance-00000023 to attach disk [datastore2] a8ceac07-6631-4ac6-b62a-7f3eac788f2b/a8ceac07-6631-4ac6-b62a-7f3eac788f2b.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 755.796070] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-979880c3-03af-4a8b-ad3b-27946a5e289a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.802305] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 755.802305] env[62753]: value = "task-1332106" [ 755.802305] env[62753]: _type = "Task" [ 755.802305] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 755.811915] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332106, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 755.916741] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1280aa6d-707f-4b87-b1c5-92f98d2a3c2c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.923774] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9f80cae-9919-4569-92e3-48c63a5fac6e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.955687] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da60c0a-02ae-483c-98c1-0bcdd7f7fa54 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.963076] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa036c11-6e64-4328-848b-eb6617f5c1c4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 755.976406] env[62753]: DEBUG nova.compute.provider_tree [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 756.011467] env[62753]: DEBUG nova.network.neutron [-] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.116785] env[62753]: DEBUG nova.network.neutron [req-7593527c-637c-464b-8754-9c99ce1f165f req-5f0d1294-8f96-446a-b274-64e8ae67535a service nova] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 756.189753] env[62753]: DEBUG nova.network.neutron [req-7593527c-637c-464b-8754-9c99ce1f165f req-5f0d1294-8f96-446a-b274-64e8ae67535a service nova] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 756.312554] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332106, 'name': Rename_Task, 'duration_secs': 0.132586} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.313877] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 756.313877] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a5f8691d-8ec5-44b6-94f2-b3e58f821d70 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.318888] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 756.318888] env[62753]: value = "task-1332107" [ 756.318888] env[62753]: _type = "Task" [ 756.318888] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 756.325995] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332107, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 756.480050] env[62753]: DEBUG nova.scheduler.client.report [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 756.513840] env[62753]: INFO nova.compute.manager [-] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Took 1.02 seconds to deallocate network for instance. [ 756.517723] env[62753]: DEBUG nova.compute.claims [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 756.517917] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 756.531481] env[62753]: DEBUG nova.compute.manager [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 756.561600] env[62753]: DEBUG nova.virt.hardware [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 756.562277] env[62753]: DEBUG nova.virt.hardware [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 756.562277] env[62753]: DEBUG nova.virt.hardware [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 756.562277] env[62753]: DEBUG nova.virt.hardware [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 756.562490] env[62753]: DEBUG nova.virt.hardware [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 756.562490] env[62753]: DEBUG nova.virt.hardware [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 756.562731] env[62753]: DEBUG nova.virt.hardware [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 756.562874] env[62753]: DEBUG nova.virt.hardware [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 756.563054] env[62753]: DEBUG nova.virt.hardware [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 756.563224] env[62753]: DEBUG nova.virt.hardware [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 756.563850] env[62753]: DEBUG nova.virt.hardware [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 756.564786] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8447a382-2d3d-4677-a0ca-d09d054f2b8a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.572363] env[62753]: ERROR nova.compute.manager [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f53e6ca3-7555-4166-b8d6-3613594ac4f8, please check neutron logs for more information. [ 756.572363] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 756.572363] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.572363] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 756.572363] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.572363] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 756.572363] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.572363] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 756.572363] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.572363] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 756.572363] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.572363] env[62753]: ERROR nova.compute.manager raise self.value [ 756.572363] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.572363] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 756.572363] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.572363] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 756.572841] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.572841] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 756.572841] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f53e6ca3-7555-4166-b8d6-3613594ac4f8, please check neutron logs for more information. [ 756.572841] env[62753]: ERROR nova.compute.manager [ 756.572841] env[62753]: Traceback (most recent call last): [ 756.572841] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 756.572841] env[62753]: listener.cb(fileno) [ 756.572841] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.572841] env[62753]: result = function(*args, **kwargs) [ 756.572841] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 756.572841] env[62753]: return func(*args, **kwargs) [ 756.572841] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.572841] env[62753]: raise e [ 756.572841] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.572841] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 756.572841] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.572841] env[62753]: created_port_ids = self._update_ports_for_instance( [ 756.572841] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.572841] env[62753]: with excutils.save_and_reraise_exception(): [ 756.572841] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.572841] env[62753]: self.force_reraise() [ 756.572841] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.572841] env[62753]: raise self.value [ 756.572841] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.572841] env[62753]: updated_port = self._update_port( [ 756.572841] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.572841] env[62753]: _ensure_no_port_binding_failure(port) [ 756.572841] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.572841] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 756.573600] env[62753]: nova.exception.PortBindingFailed: Binding failed for port f53e6ca3-7555-4166-b8d6-3613594ac4f8, please check neutron logs for more information. [ 756.573600] env[62753]: Removing descriptor: 14 [ 756.576032] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2fdc4ef8-b722-48de-8888-cefe03b8d928 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.591068] env[62753]: ERROR nova.compute.manager [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f53e6ca3-7555-4166-b8d6-3613594ac4f8, please check neutron logs for more information. [ 756.591068] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Traceback (most recent call last): [ 756.591068] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 756.591068] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] yield resources [ 756.591068] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 756.591068] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] self.driver.spawn(context, instance, image_meta, [ 756.591068] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 756.591068] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 756.591068] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 756.591068] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] vm_ref = self.build_virtual_machine(instance, [ 756.591068] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 756.591471] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] vif_infos = vmwarevif.get_vif_info(self._session, [ 756.591471] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 756.591471] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] for vif in network_info: [ 756.591471] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 756.591471] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] return self._sync_wrapper(fn, *args, **kwargs) [ 756.591471] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 756.591471] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] self.wait() [ 756.591471] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 756.591471] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] self[:] = self._gt.wait() [ 756.591471] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 756.591471] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] return self._exit_event.wait() [ 756.591471] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 756.591471] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] current.throw(*self._exc) [ 756.591914] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 756.591914] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] result = function(*args, **kwargs) [ 756.591914] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 756.591914] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] return func(*args, **kwargs) [ 756.591914] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 756.591914] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] raise e [ 756.591914] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 756.591914] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] nwinfo = self.network_api.allocate_for_instance( [ 756.591914] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 756.591914] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] created_port_ids = self._update_ports_for_instance( [ 756.591914] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 756.591914] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] with excutils.save_and_reraise_exception(): [ 756.591914] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 756.592397] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] self.force_reraise() [ 756.592397] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 756.592397] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] raise self.value [ 756.592397] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 756.592397] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] updated_port = self._update_port( [ 756.592397] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 756.592397] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] _ensure_no_port_binding_failure(port) [ 756.592397] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 756.592397] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] raise exception.PortBindingFailed(port_id=port['id']) [ 756.592397] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] nova.exception.PortBindingFailed: Binding failed for port f53e6ca3-7555-4166-b8d6-3613594ac4f8, please check neutron logs for more information. [ 756.592397] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] [ 756.592397] env[62753]: INFO nova.compute.manager [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Terminating instance [ 756.593740] env[62753]: DEBUG oslo_concurrency.lockutils [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Acquiring lock "refresh_cache-b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 756.593881] env[62753]: DEBUG oslo_concurrency.lockutils [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Acquired lock "refresh_cache-b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 756.594032] env[62753]: DEBUG nova.network.neutron [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 756.693103] env[62753]: DEBUG oslo_concurrency.lockutils [req-7593527c-637c-464b-8754-9c99ce1f165f req-5f0d1294-8f96-446a-b274-64e8ae67535a service nova] Releasing lock "refresh_cache-eca070f0-dda0-4c51-b2d3-1251c517804a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 756.693376] env[62753]: DEBUG nova.compute.manager [req-7593527c-637c-464b-8754-9c99ce1f165f req-5f0d1294-8f96-446a-b274-64e8ae67535a service nova] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Received event network-vif-deleted-b067d0d2-dcbe-4852-920e-c9130d215bdc {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 756.829122] env[62753]: DEBUG oslo_vmware.api [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332107, 'name': PowerOnVM_Task, 'duration_secs': 0.425231} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 756.829645] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 756.829645] env[62753]: DEBUG nova.compute.manager [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 756.830446] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9feb4609-2a96-4888-9253-98f2c676cf55 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 756.986888] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.480s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 756.987415] env[62753]: DEBUG nova.compute.manager [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 756.990097] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.538s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 756.991598] env[62753]: INFO nova.compute.claims [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 757.124877] env[62753]: DEBUG nova.network.neutron [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.235722] env[62753]: DEBUG nova.network.neutron [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 757.351021] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 757.496127] env[62753]: DEBUG nova.compute.utils [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 757.501893] env[62753]: DEBUG nova.compute.manager [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 757.502102] env[62753]: DEBUG nova.network.neutron [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 757.555057] env[62753]: DEBUG nova.policy [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b6207815e8041b4bba3069635d7d962', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6c8a6bc946a469fa85cc3ab80d7333a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 757.629232] env[62753]: DEBUG nova.compute.manager [req-6756943a-433b-43c5-8e30-a5a3cce1e996 req-a8a1a8cc-ffad-41eb-b1ba-497ca7c4f047 service nova] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Received event network-changed-f53e6ca3-7555-4166-b8d6-3613594ac4f8 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 757.629232] env[62753]: DEBUG nova.compute.manager [req-6756943a-433b-43c5-8e30-a5a3cce1e996 req-a8a1a8cc-ffad-41eb-b1ba-497ca7c4f047 service nova] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Refreshing instance network info cache due to event network-changed-f53e6ca3-7555-4166-b8d6-3613594ac4f8. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 757.629232] env[62753]: DEBUG oslo_concurrency.lockutils [req-6756943a-433b-43c5-8e30-a5a3cce1e996 req-a8a1a8cc-ffad-41eb-b1ba-497ca7c4f047 service nova] Acquiring lock "refresh_cache-b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 757.738411] env[62753]: DEBUG oslo_concurrency.lockutils [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Releasing lock "refresh_cache-b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 757.738889] env[62753]: DEBUG nova.compute.manager [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 757.739117] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 757.739490] env[62753]: DEBUG oslo_concurrency.lockutils [req-6756943a-433b-43c5-8e30-a5a3cce1e996 req-a8a1a8cc-ffad-41eb-b1ba-497ca7c4f047 service nova] Acquired lock "refresh_cache-b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 757.739720] env[62753]: DEBUG nova.network.neutron [req-6756943a-433b-43c5-8e30-a5a3cce1e996 req-a8a1a8cc-ffad-41eb-b1ba-497ca7c4f047 service nova] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Refreshing network info cache for port f53e6ca3-7555-4166-b8d6-3613594ac4f8 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 757.740951] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e58b80d9-701c-4f82-8d8e-fd4d65e7680c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.752473] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dfb4073a-eb91-428a-9d2a-50a21bc5840a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 757.777328] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4 could not be found. [ 757.777672] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 757.777804] env[62753]: INFO nova.compute.manager [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Took 0.04 seconds to destroy the instance on the hypervisor. [ 757.777990] env[62753]: DEBUG oslo.service.loopingcall [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 757.778219] env[62753]: DEBUG nova.compute.manager [-] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 757.778448] env[62753]: DEBUG nova.network.neutron [-] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 757.795502] env[62753]: DEBUG nova.network.neutron [-] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 757.993612] env[62753]: DEBUG nova.network.neutron [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Successfully created port: 0f17649b-6479-4040-b2cf-71fa86dbbe03 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 758.002456] env[62753]: DEBUG nova.compute.manager [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 758.272767] env[62753]: DEBUG nova.network.neutron [req-6756943a-433b-43c5-8e30-a5a3cce1e996 req-a8a1a8cc-ffad-41eb-b1ba-497ca7c4f047 service nova] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 758.298100] env[62753]: DEBUG nova.network.neutron [-] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.352389] env[62753]: DEBUG nova.network.neutron [req-6756943a-433b-43c5-8e30-a5a3cce1e996 req-a8a1a8cc-ffad-41eb-b1ba-497ca7c4f047 service nova] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 758.367637] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c8856f3-e099-41c1-a71f-da7aff87d16f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.378146] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01fbef33-f6a7-45f1-abca-3903c6a63a02 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.407982] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-67391956-a641-49d5-bee4-b3093628061a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.415007] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3855ab9b-6a80-48eb-86b0-5dd68a6ac45d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 758.428265] env[62753]: DEBUG nova.compute.provider_tree [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 758.483409] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquiring lock "a8ceac07-6631-4ac6-b62a-7f3eac788f2b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.483684] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Lock "a8ceac07-6631-4ac6-b62a-7f3eac788f2b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.484010] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquiring lock "a8ceac07-6631-4ac6-b62a-7f3eac788f2b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.484092] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Lock "a8ceac07-6631-4ac6-b62a-7f3eac788f2b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 758.484238] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Lock "a8ceac07-6631-4ac6-b62a-7f3eac788f2b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 758.487583] env[62753]: INFO nova.compute.manager [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Terminating instance [ 758.489653] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquiring lock "refresh_cache-a8ceac07-6631-4ac6-b62a-7f3eac788f2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 758.490120] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquired lock "refresh_cache-a8ceac07-6631-4ac6-b62a-7f3eac788f2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 758.490120] env[62753]: DEBUG nova.network.neutron [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 758.800626] env[62753]: INFO nova.compute.manager [-] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Took 1.02 seconds to deallocate network for instance. [ 758.803155] env[62753]: DEBUG nova.compute.claims [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 758.803366] env[62753]: DEBUG oslo_concurrency.lockutils [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 758.854596] env[62753]: DEBUG oslo_concurrency.lockutils [req-6756943a-433b-43c5-8e30-a5a3cce1e996 req-a8a1a8cc-ffad-41eb-b1ba-497ca7c4f047 service nova] Releasing lock "refresh_cache-b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 758.854840] env[62753]: DEBUG nova.compute.manager [req-6756943a-433b-43c5-8e30-a5a3cce1e996 req-a8a1a8cc-ffad-41eb-b1ba-497ca7c4f047 service nova] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Received event network-vif-deleted-f53e6ca3-7555-4166-b8d6-3613594ac4f8 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 758.931767] env[62753]: DEBUG nova.scheduler.client.report [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 759.007148] env[62753]: DEBUG nova.network.neutron [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.016343] env[62753]: DEBUG nova.compute.manager [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 759.044038] env[62753]: DEBUG nova.virt.hardware [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 759.044314] env[62753]: DEBUG nova.virt.hardware [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 759.044469] env[62753]: DEBUG nova.virt.hardware [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 759.044645] env[62753]: DEBUG nova.virt.hardware [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 759.044792] env[62753]: DEBUG nova.virt.hardware [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 759.044936] env[62753]: DEBUG nova.virt.hardware [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 759.045156] env[62753]: DEBUG nova.virt.hardware [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 759.045314] env[62753]: DEBUG nova.virt.hardware [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 759.045480] env[62753]: DEBUG nova.virt.hardware [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 759.045642] env[62753]: DEBUG nova.virt.hardware [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 759.045814] env[62753]: DEBUG nova.virt.hardware [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 759.046691] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8cdac946-cd99-454b-bfb4-a52e512afa22 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.054582] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0936b1e7-f4c0-4f17-80c3-756e7314c3e3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.069745] env[62753]: DEBUG nova.network.neutron [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.169915] env[62753]: ERROR nova.compute.manager [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 0f17649b-6479-4040-b2cf-71fa86dbbe03, please check neutron logs for more information. [ 759.169915] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 759.169915] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.169915] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 759.169915] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.169915] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 759.169915] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.169915] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 759.169915] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.169915] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 759.169915] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.169915] env[62753]: ERROR nova.compute.manager raise self.value [ 759.169915] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.169915] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 759.169915] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.169915] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 759.170455] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.170455] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 759.170455] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 0f17649b-6479-4040-b2cf-71fa86dbbe03, please check neutron logs for more information. [ 759.170455] env[62753]: ERROR nova.compute.manager [ 759.170455] env[62753]: Traceback (most recent call last): [ 759.170455] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 759.170455] env[62753]: listener.cb(fileno) [ 759.170455] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.170455] env[62753]: result = function(*args, **kwargs) [ 759.170455] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 759.170455] env[62753]: return func(*args, **kwargs) [ 759.170455] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 759.170455] env[62753]: raise e [ 759.170455] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.170455] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 759.170455] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.170455] env[62753]: created_port_ids = self._update_ports_for_instance( [ 759.170455] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.170455] env[62753]: with excutils.save_and_reraise_exception(): [ 759.170455] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.170455] env[62753]: self.force_reraise() [ 759.170455] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.170455] env[62753]: raise self.value [ 759.170455] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.170455] env[62753]: updated_port = self._update_port( [ 759.170455] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.170455] env[62753]: _ensure_no_port_binding_failure(port) [ 759.170455] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.170455] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 759.171385] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 0f17649b-6479-4040-b2cf-71fa86dbbe03, please check neutron logs for more information. [ 759.171385] env[62753]: Removing descriptor: 14 [ 759.171385] env[62753]: ERROR nova.compute.manager [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 0f17649b-6479-4040-b2cf-71fa86dbbe03, please check neutron logs for more information. [ 759.171385] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Traceback (most recent call last): [ 759.171385] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 759.171385] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] yield resources [ 759.171385] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 759.171385] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] self.driver.spawn(context, instance, image_meta, [ 759.171385] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 759.171385] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] self._vmops.spawn(context, instance, image_meta, injected_files, [ 759.171385] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 759.171385] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] vm_ref = self.build_virtual_machine(instance, [ 759.171767] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 759.171767] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] vif_infos = vmwarevif.get_vif_info(self._session, [ 759.171767] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 759.171767] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] for vif in network_info: [ 759.171767] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 759.171767] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] return self._sync_wrapper(fn, *args, **kwargs) [ 759.171767] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 759.171767] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] self.wait() [ 759.171767] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 759.171767] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] self[:] = self._gt.wait() [ 759.171767] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 759.171767] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] return self._exit_event.wait() [ 759.171767] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 759.172130] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] result = hub.switch() [ 759.172130] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 759.172130] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] return self.greenlet.switch() [ 759.172130] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 759.172130] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] result = function(*args, **kwargs) [ 759.172130] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 759.172130] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] return func(*args, **kwargs) [ 759.172130] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 759.172130] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] raise e [ 759.172130] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 759.172130] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] nwinfo = self.network_api.allocate_for_instance( [ 759.172130] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 759.172130] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] created_port_ids = self._update_ports_for_instance( [ 759.172496] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 759.172496] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] with excutils.save_and_reraise_exception(): [ 759.172496] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 759.172496] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] self.force_reraise() [ 759.172496] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 759.172496] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] raise self.value [ 759.172496] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 759.172496] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] updated_port = self._update_port( [ 759.172496] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 759.172496] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] _ensure_no_port_binding_failure(port) [ 759.172496] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 759.172496] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] raise exception.PortBindingFailed(port_id=port['id']) [ 759.172821] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] nova.exception.PortBindingFailed: Binding failed for port 0f17649b-6479-4040-b2cf-71fa86dbbe03, please check neutron logs for more information. [ 759.172821] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] [ 759.172821] env[62753]: INFO nova.compute.manager [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Terminating instance [ 759.173586] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "refresh_cache-358d1c98-28ee-4f18-bd42-89a1e7687d03" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.173760] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "refresh_cache-358d1c98-28ee-4f18-bd42-89a1e7687d03" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 759.173945] env[62753]: DEBUG nova.network.neutron [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 759.440737] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.450s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 759.441364] env[62753]: DEBUG nova.compute.manager [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 759.444125] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 12.243s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 759.573533] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Releasing lock "refresh_cache-a8ceac07-6631-4ac6-b62a-7f3eac788f2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 759.573863] env[62753]: DEBUG nova.compute.manager [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 759.574148] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 759.575035] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad93984b-bc7d-4407-9176-9723f13c6b04 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.583031] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 759.583280] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54e7bc3f-7f9b-4041-b5a9-26532574f956 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 759.588946] env[62753]: DEBUG oslo_vmware.api [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 759.588946] env[62753]: value = "task-1332108" [ 759.588946] env[62753]: _type = "Task" [ 759.588946] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 759.596547] env[62753]: DEBUG oslo_vmware.api [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332108, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 759.677010] env[62753]: DEBUG nova.compute.manager [req-e7e54d63-9f15-4216-8395-da58ac9b15b6 req-ee4a6d4a-7e53-4e9f-9a7c-2d0237530179 service nova] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Received event network-changed-0f17649b-6479-4040-b2cf-71fa86dbbe03 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 759.677249] env[62753]: DEBUG nova.compute.manager [req-e7e54d63-9f15-4216-8395-da58ac9b15b6 req-ee4a6d4a-7e53-4e9f-9a7c-2d0237530179 service nova] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Refreshing instance network info cache due to event network-changed-0f17649b-6479-4040-b2cf-71fa86dbbe03. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 759.677403] env[62753]: DEBUG oslo_concurrency.lockutils [req-e7e54d63-9f15-4216-8395-da58ac9b15b6 req-ee4a6d4a-7e53-4e9f-9a7c-2d0237530179 service nova] Acquiring lock "refresh_cache-358d1c98-28ee-4f18-bd42-89a1e7687d03" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 759.696766] env[62753]: DEBUG nova.network.neutron [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 759.802798] env[62753]: DEBUG nova.network.neutron [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 759.951461] env[62753]: DEBUG nova.compute.utils [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 759.953068] env[62753]: DEBUG nova.compute.manager [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 759.953179] env[62753]: DEBUG nova.network.neutron [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 760.099080] env[62753]: DEBUG oslo_vmware.api [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332108, 'name': PowerOffVM_Task, 'duration_secs': 0.191981} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.101240] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 760.101412] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 760.102118] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5ad828a8-b259-4ff0-a37a-13fc046797b8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.128918] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 760.129142] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 760.129331] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Deleting the datastore file [datastore2] a8ceac07-6631-4ac6-b62a-7f3eac788f2b {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 760.129650] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-95241dd0-32d1-4fe6-80c8-dda70ebdbb71 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.137963] env[62753]: DEBUG oslo_vmware.api [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for the task: (returnval){ [ 760.137963] env[62753]: value = "task-1332110" [ 760.137963] env[62753]: _type = "Task" [ 760.137963] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 760.145896] env[62753]: DEBUG oslo_vmware.api [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332110, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 760.157776] env[62753]: DEBUG nova.policy [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '08b74083fa7248e7bd544fe7fecee7e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d0eb0b3ba504f59a85cb021da2b47b2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 760.248850] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cb81e5b-e309-4033-944d-bdedd2badc34 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.256148] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4c94f7b-cbaf-4f14-a114-65cc96a72e85 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.288244] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcbe6a68-417e-4654-9f4c-11732fd82e73 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.295572] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c64a45c-8fb1-4a98-8e36-ce4b07248012 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.308627] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "refresh_cache-358d1c98-28ee-4f18-bd42-89a1e7687d03" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 760.309017] env[62753]: DEBUG nova.compute.manager [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 760.309212] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 760.309697] env[62753]: DEBUG nova.compute.provider_tree [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 760.310998] env[62753]: DEBUG oslo_concurrency.lockutils [req-e7e54d63-9f15-4216-8395-da58ac9b15b6 req-ee4a6d4a-7e53-4e9f-9a7c-2d0237530179 service nova] Acquired lock "refresh_cache-358d1c98-28ee-4f18-bd42-89a1e7687d03" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 760.314272] env[62753]: DEBUG nova.network.neutron [req-e7e54d63-9f15-4216-8395-da58ac9b15b6 req-ee4a6d4a-7e53-4e9f-9a7c-2d0237530179 service nova] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Refreshing network info cache for port 0f17649b-6479-4040-b2cf-71fa86dbbe03 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 760.314272] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-adec6e64-2800-48a5-850a-99fbfee7cef6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.319626] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30457c1b-f835-4f6d-b57f-70853b0bd018 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 760.340952] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 358d1c98-28ee-4f18-bd42-89a1e7687d03 could not be found. [ 760.341211] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 760.341393] env[62753]: INFO nova.compute.manager [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Took 0.03 seconds to destroy the instance on the hypervisor. [ 760.341658] env[62753]: DEBUG oslo.service.loopingcall [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 760.341878] env[62753]: DEBUG nova.compute.manager [-] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 760.341971] env[62753]: DEBUG nova.network.neutron [-] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 760.358380] env[62753]: DEBUG nova.network.neutron [-] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.456364] env[62753]: DEBUG nova.compute.manager [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 760.535098] env[62753]: DEBUG nova.network.neutron [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Successfully created port: 7d800493-1971-4372-8c86-b9d4ff0be887 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 760.647256] env[62753]: DEBUG oslo_vmware.api [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Task: {'id': task-1332110, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.106226} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 760.647509] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 760.647689] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 760.647861] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 760.648041] env[62753]: INFO nova.compute.manager [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Took 1.07 seconds to destroy the instance on the hypervisor. [ 760.648280] env[62753]: DEBUG oslo.service.loopingcall [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 760.648624] env[62753]: DEBUG nova.compute.manager [-] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 760.648624] env[62753]: DEBUG nova.network.neutron [-] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 760.664045] env[62753]: DEBUG nova.network.neutron [-] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.815230] env[62753]: DEBUG nova.scheduler.client.report [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 760.839898] env[62753]: DEBUG nova.network.neutron [req-e7e54d63-9f15-4216-8395-da58ac9b15b6 req-ee4a6d4a-7e53-4e9f-9a7c-2d0237530179 service nova] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 760.860632] env[62753]: DEBUG nova.network.neutron [-] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 760.930557] env[62753]: DEBUG nova.network.neutron [req-e7e54d63-9f15-4216-8395-da58ac9b15b6 req-ee4a6d4a-7e53-4e9f-9a7c-2d0237530179 service nova] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.166621] env[62753]: DEBUG nova.network.neutron [-] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 761.321414] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.877s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 761.322093] env[62753]: ERROR nova.compute.manager [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port bd51ad15-871e-4a77-bdea-d828df308468, please check neutron logs for more information. [ 761.322093] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Traceback (most recent call last): [ 761.322093] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 761.322093] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] self.driver.spawn(context, instance, image_meta, [ 761.322093] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 761.322093] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] self._vmops.spawn(context, instance, image_meta, injected_files, [ 761.322093] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 761.322093] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] vm_ref = self.build_virtual_machine(instance, [ 761.322093] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 761.322093] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] vif_infos = vmwarevif.get_vif_info(self._session, [ 761.322093] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 761.322469] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] for vif in network_info: [ 761.322469] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 761.322469] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] return self._sync_wrapper(fn, *args, **kwargs) [ 761.322469] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 761.322469] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] self.wait() [ 761.322469] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 761.322469] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] self[:] = self._gt.wait() [ 761.322469] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 761.322469] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] return self._exit_event.wait() [ 761.322469] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 761.322469] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] result = hub.switch() [ 761.322469] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 761.322469] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] return self.greenlet.switch() [ 761.322810] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.322810] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] result = function(*args, **kwargs) [ 761.322810] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 761.322810] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] return func(*args, **kwargs) [ 761.322810] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.322810] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] raise e [ 761.322810] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.322810] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] nwinfo = self.network_api.allocate_for_instance( [ 761.322810] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.322810] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] created_port_ids = self._update_ports_for_instance( [ 761.322810] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.322810] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] with excutils.save_and_reraise_exception(): [ 761.322810] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.323163] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] self.force_reraise() [ 761.323163] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.323163] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] raise self.value [ 761.323163] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.323163] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] updated_port = self._update_port( [ 761.323163] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.323163] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] _ensure_no_port_binding_failure(port) [ 761.323163] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.323163] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] raise exception.PortBindingFailed(port_id=port['id']) [ 761.323163] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] nova.exception.PortBindingFailed: Binding failed for port bd51ad15-871e-4a77-bdea-d828df308468, please check neutron logs for more information. [ 761.323163] env[62753]: ERROR nova.compute.manager [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] [ 761.323443] env[62753]: DEBUG nova.compute.utils [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Binding failed for port bd51ad15-871e-4a77-bdea-d828df308468, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 761.324080] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.346s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 761.326096] env[62753]: INFO nova.compute.claims [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 761.328552] env[62753]: DEBUG nova.compute.manager [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Build of instance da9b45e9-618d-4e1e-9c49-3d702b2435e9 was re-scheduled: Binding failed for port bd51ad15-871e-4a77-bdea-d828df308468, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 761.329021] env[62753]: DEBUG nova.compute.manager [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 761.329285] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Acquiring lock "refresh_cache-da9b45e9-618d-4e1e-9c49-3d702b2435e9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.329469] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Acquired lock "refresh_cache-da9b45e9-618d-4e1e-9c49-3d702b2435e9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.329681] env[62753]: DEBUG nova.network.neutron [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 761.365571] env[62753]: INFO nova.compute.manager [-] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Took 1.02 seconds to deallocate network for instance. [ 761.367702] env[62753]: DEBUG nova.compute.claims [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 761.367879] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 761.432885] env[62753]: DEBUG oslo_concurrency.lockutils [req-e7e54d63-9f15-4216-8395-da58ac9b15b6 req-ee4a6d4a-7e53-4e9f-9a7c-2d0237530179 service nova] Releasing lock "refresh_cache-358d1c98-28ee-4f18-bd42-89a1e7687d03" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 761.433103] env[62753]: DEBUG nova.compute.manager [req-e7e54d63-9f15-4216-8395-da58ac9b15b6 req-ee4a6d4a-7e53-4e9f-9a7c-2d0237530179 service nova] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Received event network-vif-deleted-0f17649b-6479-4040-b2cf-71fa86dbbe03 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.468772] env[62753]: DEBUG nova.compute.manager [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 761.499589] env[62753]: DEBUG nova.virt.hardware [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 761.499736] env[62753]: DEBUG nova.virt.hardware [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 761.499782] env[62753]: DEBUG nova.virt.hardware [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 761.499965] env[62753]: DEBUG nova.virt.hardware [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 761.500123] env[62753]: DEBUG nova.virt.hardware [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 761.500270] env[62753]: DEBUG nova.virt.hardware [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 761.500474] env[62753]: DEBUG nova.virt.hardware [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 761.500660] env[62753]: DEBUG nova.virt.hardware [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 761.500796] env[62753]: DEBUG nova.virt.hardware [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 761.500952] env[62753]: DEBUG nova.virt.hardware [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 761.501309] env[62753]: DEBUG nova.virt.hardware [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 761.502217] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c4a9a01-e6e6-48fa-bac3-b8845fc9f6d9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.509965] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9e1303b-c526-4bb3-b000-8209bdb18a9d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 761.626823] env[62753]: ERROR nova.compute.manager [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 7d800493-1971-4372-8c86-b9d4ff0be887, please check neutron logs for more information. [ 761.626823] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 761.626823] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.626823] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 761.626823] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.626823] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 761.626823] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.626823] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 761.626823] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.626823] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 761.626823] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.626823] env[62753]: ERROR nova.compute.manager raise self.value [ 761.626823] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.626823] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 761.626823] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.626823] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 761.627284] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.627284] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 761.627284] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 7d800493-1971-4372-8c86-b9d4ff0be887, please check neutron logs for more information. [ 761.627284] env[62753]: ERROR nova.compute.manager [ 761.627284] env[62753]: Traceback (most recent call last): [ 761.627284] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 761.627284] env[62753]: listener.cb(fileno) [ 761.627284] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.627284] env[62753]: result = function(*args, **kwargs) [ 761.627284] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 761.627284] env[62753]: return func(*args, **kwargs) [ 761.627284] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.627284] env[62753]: raise e [ 761.627284] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.627284] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 761.627284] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.627284] env[62753]: created_port_ids = self._update_ports_for_instance( [ 761.627284] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.627284] env[62753]: with excutils.save_and_reraise_exception(): [ 761.627284] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.627284] env[62753]: self.force_reraise() [ 761.627284] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.627284] env[62753]: raise self.value [ 761.627284] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.627284] env[62753]: updated_port = self._update_port( [ 761.627284] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.627284] env[62753]: _ensure_no_port_binding_failure(port) [ 761.627284] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.627284] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 761.628084] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 7d800493-1971-4372-8c86-b9d4ff0be887, please check neutron logs for more information. [ 761.628084] env[62753]: Removing descriptor: 14 [ 761.628084] env[62753]: ERROR nova.compute.manager [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 7d800493-1971-4372-8c86-b9d4ff0be887, please check neutron logs for more information. [ 761.628084] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Traceback (most recent call last): [ 761.628084] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 761.628084] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] yield resources [ 761.628084] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 761.628084] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] self.driver.spawn(context, instance, image_meta, [ 761.628084] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 761.628084] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] self._vmops.spawn(context, instance, image_meta, injected_files, [ 761.628084] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 761.628084] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] vm_ref = self.build_virtual_machine(instance, [ 761.628462] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 761.628462] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] vif_infos = vmwarevif.get_vif_info(self._session, [ 761.628462] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 761.628462] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] for vif in network_info: [ 761.628462] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 761.628462] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] return self._sync_wrapper(fn, *args, **kwargs) [ 761.628462] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 761.628462] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] self.wait() [ 761.628462] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 761.628462] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] self[:] = self._gt.wait() [ 761.628462] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 761.628462] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] return self._exit_event.wait() [ 761.628462] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 761.628863] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] result = hub.switch() [ 761.628863] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 761.628863] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] return self.greenlet.switch() [ 761.628863] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 761.628863] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] result = function(*args, **kwargs) [ 761.628863] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 761.628863] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] return func(*args, **kwargs) [ 761.628863] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 761.628863] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] raise e [ 761.628863] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 761.628863] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] nwinfo = self.network_api.allocate_for_instance( [ 761.628863] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 761.628863] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] created_port_ids = self._update_ports_for_instance( [ 761.629256] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 761.629256] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] with excutils.save_and_reraise_exception(): [ 761.629256] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 761.629256] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] self.force_reraise() [ 761.629256] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 761.629256] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] raise self.value [ 761.629256] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 761.629256] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] updated_port = self._update_port( [ 761.629256] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 761.629256] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] _ensure_no_port_binding_failure(port) [ 761.629256] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 761.629256] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] raise exception.PortBindingFailed(port_id=port['id']) [ 761.629618] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] nova.exception.PortBindingFailed: Binding failed for port 7d800493-1971-4372-8c86-b9d4ff0be887, please check neutron logs for more information. [ 761.629618] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] [ 761.629618] env[62753]: INFO nova.compute.manager [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Terminating instance [ 761.629961] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "refresh_cache-365fe802-f5e1-47d6-807e-a3b1027def02" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.630136] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquired lock "refresh_cache-365fe802-f5e1-47d6-807e-a3b1027def02" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 761.630297] env[62753]: DEBUG nova.network.neutron [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 761.673301] env[62753]: INFO nova.compute.manager [-] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Took 1.02 seconds to deallocate network for instance. [ 761.744254] env[62753]: DEBUG nova.compute.manager [req-b3a672b5-c0d1-4813-90cb-c41132be0e10 req-9183ff2f-a46e-46f0-b508-0f32c2cf1252 service nova] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Received event network-changed-7d800493-1971-4372-8c86-b9d4ff0be887 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 761.744445] env[62753]: DEBUG nova.compute.manager [req-b3a672b5-c0d1-4813-90cb-c41132be0e10 req-9183ff2f-a46e-46f0-b508-0f32c2cf1252 service nova] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Refreshing instance network info cache due to event network-changed-7d800493-1971-4372-8c86-b9d4ff0be887. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 761.744639] env[62753]: DEBUG oslo_concurrency.lockutils [req-b3a672b5-c0d1-4813-90cb-c41132be0e10 req-9183ff2f-a46e-46f0-b508-0f32c2cf1252 service nova] Acquiring lock "refresh_cache-365fe802-f5e1-47d6-807e-a3b1027def02" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 761.854921] env[62753]: DEBUG nova.network.neutron [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 761.967194] env[62753]: DEBUG nova.network.neutron [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.147108] env[62753]: DEBUG nova.network.neutron [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.180037] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 762.214586] env[62753]: DEBUG nova.network.neutron [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 762.469844] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Releasing lock "refresh_cache-da9b45e9-618d-4e1e-9c49-3d702b2435e9" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.470091] env[62753]: DEBUG nova.compute.manager [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 762.470265] env[62753]: DEBUG nova.compute.manager [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.470633] env[62753]: DEBUG nova.network.neutron [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 762.486584] env[62753]: DEBUG nova.network.neutron [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.622295] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77105165-3273-4f22-8fba-7333b5134a03 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.630046] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47cb5b9c-978a-408e-bf9d-a38d55dd5991 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.660890] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68642f2b-8b9d-4cce-8177-07fcab08eadf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.667849] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63fc3b3c-ea15-45be-8d28-ceb458d81352 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.680483] env[62753]: DEBUG nova.compute.provider_tree [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 762.717827] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Releasing lock "refresh_cache-365fe802-f5e1-47d6-807e-a3b1027def02" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 762.718263] env[62753]: DEBUG nova.compute.manager [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 762.718449] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 762.718735] env[62753]: DEBUG oslo_concurrency.lockutils [req-b3a672b5-c0d1-4813-90cb-c41132be0e10 req-9183ff2f-a46e-46f0-b508-0f32c2cf1252 service nova] Acquired lock "refresh_cache-365fe802-f5e1-47d6-807e-a3b1027def02" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 762.718907] env[62753]: DEBUG nova.network.neutron [req-b3a672b5-c0d1-4813-90cb-c41132be0e10 req-9183ff2f-a46e-46f0-b508-0f32c2cf1252 service nova] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Refreshing network info cache for port 7d800493-1971-4372-8c86-b9d4ff0be887 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 762.719864] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ec7f2701-d51c-45e3-96ef-ea216f19de42 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.729019] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07897795-69b9-48a6-9f36-ea4fa25093ce {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 762.749596] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 365fe802-f5e1-47d6-807e-a3b1027def02 could not be found. [ 762.749838] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 762.750039] env[62753]: INFO nova.compute.manager [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Took 0.03 seconds to destroy the instance on the hypervisor. [ 762.750274] env[62753]: DEBUG oslo.service.loopingcall [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 762.750479] env[62753]: DEBUG nova.compute.manager [-] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 762.750575] env[62753]: DEBUG nova.network.neutron [-] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 762.773362] env[62753]: DEBUG nova.network.neutron [-] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 762.989656] env[62753]: DEBUG nova.network.neutron [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.183467] env[62753]: DEBUG nova.scheduler.client.report [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 763.238489] env[62753]: DEBUG nova.network.neutron [req-b3a672b5-c0d1-4813-90cb-c41132be0e10 req-9183ff2f-a46e-46f0-b508-0f32c2cf1252 service nova] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 763.275186] env[62753]: DEBUG nova.network.neutron [-] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.320047] env[62753]: DEBUG nova.network.neutron [req-b3a672b5-c0d1-4813-90cb-c41132be0e10 req-9183ff2f-a46e-46f0-b508-0f32c2cf1252 service nova] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 763.492519] env[62753]: INFO nova.compute.manager [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] [instance: da9b45e9-618d-4e1e-9c49-3d702b2435e9] Took 1.02 seconds to deallocate network for instance. [ 763.689049] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.364s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 763.689049] env[62753]: DEBUG nova.compute.manager [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 763.691420] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 13.962s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 763.692959] env[62753]: INFO nova.compute.claims [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 763.777665] env[62753]: INFO nova.compute.manager [-] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Took 1.03 seconds to deallocate network for instance. [ 763.780114] env[62753]: DEBUG nova.compute.claims [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 763.780114] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 763.822037] env[62753]: DEBUG oslo_concurrency.lockutils [req-b3a672b5-c0d1-4813-90cb-c41132be0e10 req-9183ff2f-a46e-46f0-b508-0f32c2cf1252 service nova] Releasing lock "refresh_cache-365fe802-f5e1-47d6-807e-a3b1027def02" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 763.822308] env[62753]: DEBUG nova.compute.manager [req-b3a672b5-c0d1-4813-90cb-c41132be0e10 req-9183ff2f-a46e-46f0-b508-0f32c2cf1252 service nova] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Received event network-vif-deleted-7d800493-1971-4372-8c86-b9d4ff0be887 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 764.200277] env[62753]: DEBUG nova.compute.utils [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 764.201589] env[62753]: DEBUG nova.compute.manager [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Not allocating networking since 'none' was specified. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 764.519567] env[62753]: INFO nova.scheduler.client.report [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Deleted allocations for instance da9b45e9-618d-4e1e-9c49-3d702b2435e9 [ 764.705517] env[62753]: DEBUG nova.compute.manager [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 764.960781] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad35563f-158f-4558-b00a-bdd49fbd68e2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.968262] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-010bdd19-2fb6-4e95-beca-258236a97cb9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 764.997055] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cb8bde2-7b7b-4eef-a6a1-e258e2777e0f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.004370] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98ac54c2-ea80-4690-91b4-e35127a2aca6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.018204] env[62753]: DEBUG nova.compute.provider_tree [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 765.027194] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fceeb292-1ba0-4b56-a2dd-d0a9fe32fbed tempest-VolumesAdminNegativeTest-752176088 tempest-VolumesAdminNegativeTest-752176088-project-member] Lock "da9b45e9-618d-4e1e-9c49-3d702b2435e9" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 127.666s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 765.155036] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.155036] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.522834] env[62753]: DEBUG nova.scheduler.client.report [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 765.529506] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 765.662217] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 765.662217] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Starting heal instance info cache {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 765.662217] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Rebuilding the list of instances to heal {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 765.714189] env[62753]: DEBUG nova.compute.manager [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 765.747033] env[62753]: DEBUG nova.virt.hardware [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 765.747287] env[62753]: DEBUG nova.virt.hardware [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 765.747443] env[62753]: DEBUG nova.virt.hardware [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 765.748362] env[62753]: DEBUG nova.virt.hardware [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 765.748362] env[62753]: DEBUG nova.virt.hardware [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 765.748362] env[62753]: DEBUG nova.virt.hardware [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 765.748362] env[62753]: DEBUG nova.virt.hardware [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 765.748362] env[62753]: DEBUG nova.virt.hardware [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 765.748770] env[62753]: DEBUG nova.virt.hardware [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 765.748770] env[62753]: DEBUG nova.virt.hardware [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 765.748770] env[62753]: DEBUG nova.virt.hardware [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 765.749592] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49273350-8a2f-4ee0-82e2-47fc9142380d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.758557] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea1b6a2e-cf87-4725-b900-f244b09d51ff {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.772000] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Instance VIF info [] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 765.778819] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Creating folder: Project (8df18668407645c5a414c66a05fae498). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 765.778819] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ddb3778a-4727-4dc4-a1db-b57c8de4b2e3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.787853] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Created folder: Project (8df18668407645c5a414c66a05fae498) in parent group-v284541. [ 765.788039] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Creating folder: Instances. Parent ref: group-v284549. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 765.788275] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1f04b048-f840-4954-a305-8c594cfd24d0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.796615] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Created folder: Instances in parent group-v284549. [ 765.797910] env[62753]: DEBUG oslo.service.loopingcall [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 765.797910] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 765.797910] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2105b5dd-f756-4024-9016-d5794b71320a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 765.813749] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 765.813749] env[62753]: value = "task-1332113" [ 765.813749] env[62753]: _type = "Task" [ 765.813749] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 765.820847] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332113, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.033032] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.341s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 766.033704] env[62753]: DEBUG nova.compute.manager [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 766.036866] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.059s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 766.038753] env[62753]: INFO nova.compute.claims [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 766.070460] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 766.165253] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 766.165823] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 766.165823] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 766.165823] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 766.165823] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 766.165999] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 766.165999] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 766.208677] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "refresh_cache-a8ceac07-6631-4ac6-b62a-7f3eac788f2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.208832] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquired lock "refresh_cache-a8ceac07-6631-4ac6-b62a-7f3eac788f2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.208979] env[62753]: DEBUG nova.network.neutron [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Forcefully refreshing network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 766.209196] env[62753]: DEBUG nova.objects.instance [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lazy-loading 'info_cache' on Instance uuid a8ceac07-6631-4ac6-b62a-7f3eac788f2b {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 766.324987] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332113, 'name': CreateVM_Task, 'duration_secs': 0.249175} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.325178] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 766.325590] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.325746] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.326071] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 766.326317] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1e43ec8e-8c75-44f7-9607-6e51aee23925 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.330644] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Waiting for the task: (returnval){ [ 766.330644] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52effb96-85c8-ead6-137a-81701f567383" [ 766.330644] env[62753]: _type = "Task" [ 766.330644] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.337958] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52effb96-85c8-ead6-137a-81701f567383, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 766.551685] env[62753]: DEBUG nova.compute.utils [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 766.558713] env[62753]: DEBUG nova.compute.manager [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 766.558713] env[62753]: DEBUG nova.network.neutron [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 766.610442] env[62753]: DEBUG nova.policy [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e59df77fc9bd49788aa7351eb53f80ec', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'ad0eda57d7b847d8b298ecbc7deb8bad', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 766.840927] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52effb96-85c8-ead6-137a-81701f567383, 'name': SearchDatastore_Task, 'duration_secs': 0.0084} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 766.841315] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 766.841566] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 766.841887] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 766.842041] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 766.842225] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 766.842472] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-14527970-9475-4f28-a1fe-1e8f34639836 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.849986] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 766.850177] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 766.850907] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a58070f4-6f74-49b9-a2fe-e93581d4a005 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 766.855770] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Waiting for the task: (returnval){ [ 766.855770] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5231a69b-7eef-1559-b890-3771b508d46e" [ 766.855770] env[62753]: _type = "Task" [ 766.855770] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 766.862793] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5231a69b-7eef-1559-b890-3771b508d46e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.019909] env[62753]: DEBUG nova.network.neutron [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Successfully created port: 3df0753e-2da2-4367-9024-eac571d2ebd3 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 767.062603] env[62753]: DEBUG nova.compute.manager [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 767.233368] env[62753]: DEBUG nova.network.neutron [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 767.372473] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5231a69b-7eef-1559-b890-3771b508d46e, 'name': SearchDatastore_Task, 'duration_secs': 0.008043} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.378259] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4871f336-2a4e-415a-b25d-bae265e0a4e8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.384137] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Waiting for the task: (returnval){ [ 767.384137] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52b51aff-cea5-8d95-662e-ef5a8d7d51b7" [ 767.384137] env[62753]: _type = "Task" [ 767.384137] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.395654] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52b51aff-cea5-8d95-662e-ef5a8d7d51b7, 'name': SearchDatastore_Task, 'duration_secs': 0.008113} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.396039] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 767.396301] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 8664ed5c-ecb9-4795-8499-31198cfd0450/8664ed5c-ecb9-4795-8499-31198cfd0450.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 767.396537] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-cde4f743-1e73-4383-a7ca-006d3c9bd4f6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.400065] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c1504f1-8133-4894-b5f7-6f0cfa2205b0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.409016] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-368ae6d6-678d-4366-b862-1723557cce2e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.413720] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Waiting for the task: (returnval){ [ 767.413720] env[62753]: value = "task-1332114" [ 767.413720] env[62753]: _type = "Task" [ 767.413720] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.444013] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-574cf5ac-c744-4838-9c99-c5b32e0893b5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.449561] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': task-1332114, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.454363] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d31a67f3-859e-40c4-9463-04fd0c6a9e23 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.468808] env[62753]: DEBUG nova.compute.provider_tree [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 767.880287] env[62753]: DEBUG nova.network.neutron [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 767.921352] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': task-1332114, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.471708} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 767.924628] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 8664ed5c-ecb9-4795-8499-31198cfd0450/8664ed5c-ecb9-4795-8499-31198cfd0450.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 767.924628] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 767.924628] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ac780c58-c37d-4b3d-8ed8-fd1c28208bf4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 767.929106] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Waiting for the task: (returnval){ [ 767.929106] env[62753]: value = "task-1332115" [ 767.929106] env[62753]: _type = "Task" [ 767.929106] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 767.938223] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': task-1332115, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 767.972858] env[62753]: DEBUG nova.scheduler.client.report [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 768.059931] env[62753]: DEBUG nova.compute.manager [req-c5eeba6e-0732-4227-9c88-d1d69fcfc6c8 req-75564b04-859d-4239-98c1-3459fab5f472 service nova] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Received event network-changed-3df0753e-2da2-4367-9024-eac571d2ebd3 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 768.059931] env[62753]: DEBUG nova.compute.manager [req-c5eeba6e-0732-4227-9c88-d1d69fcfc6c8 req-75564b04-859d-4239-98c1-3459fab5f472 service nova] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Refreshing instance network info cache due to event network-changed-3df0753e-2da2-4367-9024-eac571d2ebd3. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 768.059931] env[62753]: DEBUG oslo_concurrency.lockutils [req-c5eeba6e-0732-4227-9c88-d1d69fcfc6c8 req-75564b04-859d-4239-98c1-3459fab5f472 service nova] Acquiring lock "refresh_cache-73f7840f-6e3e-4a4e-b7f5-bffe3009aae6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.060349] env[62753]: DEBUG oslo_concurrency.lockutils [req-c5eeba6e-0732-4227-9c88-d1d69fcfc6c8 req-75564b04-859d-4239-98c1-3459fab5f472 service nova] Acquired lock "refresh_cache-73f7840f-6e3e-4a4e-b7f5-bffe3009aae6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 768.060349] env[62753]: DEBUG nova.network.neutron [req-c5eeba6e-0732-4227-9c88-d1d69fcfc6c8 req-75564b04-859d-4239-98c1-3459fab5f472 service nova] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Refreshing network info cache for port 3df0753e-2da2-4367-9024-eac571d2ebd3 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 768.076203] env[62753]: DEBUG nova.compute.manager [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 768.106990] env[62753]: DEBUG nova.virt.hardware [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 768.107275] env[62753]: DEBUG nova.virt.hardware [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 768.107759] env[62753]: DEBUG nova.virt.hardware [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 768.107759] env[62753]: DEBUG nova.virt.hardware [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 768.107759] env[62753]: DEBUG nova.virt.hardware [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 768.107916] env[62753]: DEBUG nova.virt.hardware [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 768.108649] env[62753]: DEBUG nova.virt.hardware [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 768.108649] env[62753]: DEBUG nova.virt.hardware [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 768.108649] env[62753]: DEBUG nova.virt.hardware [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 768.108649] env[62753]: DEBUG nova.virt.hardware [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 768.108848] env[62753]: DEBUG nova.virt.hardware [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 768.109659] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d65b69e-b509-437b-a440-d7b95357a4f8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.118228] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-812f3876-c4a6-4b55-9bc5-4ecf19335c82 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.228622] env[62753]: ERROR nova.compute.manager [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 3df0753e-2da2-4367-9024-eac571d2ebd3, please check neutron logs for more information. [ 768.228622] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 768.228622] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 768.228622] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 768.228622] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 768.228622] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 768.228622] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 768.228622] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 768.228622] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 768.228622] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 768.228622] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 768.228622] env[62753]: ERROR nova.compute.manager raise self.value [ 768.228622] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 768.228622] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 768.228622] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 768.228622] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 768.229298] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 768.229298] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 768.229298] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 3df0753e-2da2-4367-9024-eac571d2ebd3, please check neutron logs for more information. [ 768.229298] env[62753]: ERROR nova.compute.manager [ 768.229298] env[62753]: Traceback (most recent call last): [ 768.229298] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 768.229298] env[62753]: listener.cb(fileno) [ 768.229298] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 768.229298] env[62753]: result = function(*args, **kwargs) [ 768.229298] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 768.229298] env[62753]: return func(*args, **kwargs) [ 768.229298] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 768.229298] env[62753]: raise e [ 768.229298] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 768.229298] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 768.229298] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 768.229298] env[62753]: created_port_ids = self._update_ports_for_instance( [ 768.229298] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 768.229298] env[62753]: with excutils.save_and_reraise_exception(): [ 768.229298] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 768.229298] env[62753]: self.force_reraise() [ 768.229298] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 768.229298] env[62753]: raise self.value [ 768.229298] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 768.229298] env[62753]: updated_port = self._update_port( [ 768.229298] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 768.229298] env[62753]: _ensure_no_port_binding_failure(port) [ 768.229298] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 768.229298] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 768.230071] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 3df0753e-2da2-4367-9024-eac571d2ebd3, please check neutron logs for more information. [ 768.230071] env[62753]: Removing descriptor: 14 [ 768.230071] env[62753]: ERROR nova.compute.manager [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 3df0753e-2da2-4367-9024-eac571d2ebd3, please check neutron logs for more information. [ 768.230071] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Traceback (most recent call last): [ 768.230071] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 768.230071] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] yield resources [ 768.230071] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 768.230071] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] self.driver.spawn(context, instance, image_meta, [ 768.230071] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 768.230071] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 768.230071] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 768.230071] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] vm_ref = self.build_virtual_machine(instance, [ 768.230399] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 768.230399] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] vif_infos = vmwarevif.get_vif_info(self._session, [ 768.230399] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 768.230399] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] for vif in network_info: [ 768.230399] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 768.230399] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] return self._sync_wrapper(fn, *args, **kwargs) [ 768.230399] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 768.230399] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] self.wait() [ 768.230399] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 768.230399] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] self[:] = self._gt.wait() [ 768.230399] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 768.230399] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] return self._exit_event.wait() [ 768.230399] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 768.230748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] result = hub.switch() [ 768.230748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 768.230748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] return self.greenlet.switch() [ 768.230748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 768.230748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] result = function(*args, **kwargs) [ 768.230748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 768.230748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] return func(*args, **kwargs) [ 768.230748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 768.230748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] raise e [ 768.230748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 768.230748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] nwinfo = self.network_api.allocate_for_instance( [ 768.230748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 768.230748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] created_port_ids = self._update_ports_for_instance( [ 768.231099] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 768.231099] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] with excutils.save_and_reraise_exception(): [ 768.231099] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 768.231099] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] self.force_reraise() [ 768.231099] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 768.231099] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] raise self.value [ 768.231099] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 768.231099] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] updated_port = self._update_port( [ 768.231099] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 768.231099] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] _ensure_no_port_binding_failure(port) [ 768.231099] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 768.231099] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] raise exception.PortBindingFailed(port_id=port['id']) [ 768.231484] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] nova.exception.PortBindingFailed: Binding failed for port 3df0753e-2da2-4367-9024-eac571d2ebd3, please check neutron logs for more information. [ 768.231484] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] [ 768.231484] env[62753]: INFO nova.compute.manager [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Terminating instance [ 768.231991] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Acquiring lock "refresh_cache-73f7840f-6e3e-4a4e-b7f5-bffe3009aae6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 768.383871] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Releasing lock "refresh_cache-a8ceac07-6631-4ac6-b62a-7f3eac788f2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 768.383871] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Updated the network info_cache for instance {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 768.383993] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.384178] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.384329] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.384475] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.384615] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.384755] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.384883] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62753) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 768.385034] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 768.439059] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': task-1332115, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.053874} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 768.439339] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 768.440127] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0fb358d-00a7-4f30-97a1-4b481c6b5e1d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.464535] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Reconfiguring VM instance instance-0000002a to attach disk [datastore2] 8664ed5c-ecb9-4795-8499-31198cfd0450/8664ed5c-ecb9-4795-8499-31198cfd0450.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 768.465185] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a0e352ae-b06a-4514-8519-fa6c39fc592b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 768.480038] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.443s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 768.480456] env[62753]: DEBUG nova.compute.manager [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 768.483572] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.565s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 768.488015] env[62753]: INFO nova.compute.claims [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 768.492078] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Waiting for the task: (returnval){ [ 768.492078] env[62753]: value = "task-1332116" [ 768.492078] env[62753]: _type = "Task" [ 768.492078] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 768.501229] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': task-1332116, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 768.581366] env[62753]: DEBUG nova.network.neutron [req-c5eeba6e-0732-4227-9c88-d1d69fcfc6c8 req-75564b04-859d-4239-98c1-3459fab5f472 service nova] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 768.739810] env[62753]: DEBUG nova.network.neutron [req-c5eeba6e-0732-4227-9c88-d1d69fcfc6c8 req-75564b04-859d-4239-98c1-3459fab5f472 service nova] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 768.888537] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 768.989228] env[62753]: DEBUG nova.compute.utils [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 768.993175] env[62753]: DEBUG nova.compute.manager [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 768.993363] env[62753]: DEBUG nova.network.neutron [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 769.005128] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': task-1332116, 'name': ReconfigVM_Task, 'duration_secs': 0.259475} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.006268] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Reconfigured VM instance instance-0000002a to attach disk [datastore2] 8664ed5c-ecb9-4795-8499-31198cfd0450/8664ed5c-ecb9-4795-8499-31198cfd0450.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 769.006802] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-dc75e29a-df56-4167-ada0-234251e02d3d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.013423] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Waiting for the task: (returnval){ [ 769.013423] env[62753]: value = "task-1332117" [ 769.013423] env[62753]: _type = "Task" [ 769.013423] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.021763] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': task-1332117, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.064771] env[62753]: DEBUG nova.policy [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fe29a17b9b2243d98875708aeac08637', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'd0997fb64a304242b3ce2df58fc0df53', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 769.243239] env[62753]: DEBUG oslo_concurrency.lockutils [req-c5eeba6e-0732-4227-9c88-d1d69fcfc6c8 req-75564b04-859d-4239-98c1-3459fab5f472 service nova] Releasing lock "refresh_cache-73f7840f-6e3e-4a4e-b7f5-bffe3009aae6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 769.244218] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Acquired lock "refresh_cache-73f7840f-6e3e-4a4e-b7f5-bffe3009aae6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 769.244218] env[62753]: DEBUG nova.network.neutron [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 769.411938] env[62753]: DEBUG nova.network.neutron [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Successfully created port: b2ddb78b-e696-4141-8703-b25192d9afaa {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 769.496718] env[62753]: DEBUG nova.compute.manager [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 769.534244] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': task-1332117, 'name': Rename_Task, 'duration_secs': 0.127105} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 769.534724] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 769.534986] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a358d428-dca2-4be8-b185-81823b210e92 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.541765] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Waiting for the task: (returnval){ [ 769.541765] env[62753]: value = "task-1332118" [ 769.541765] env[62753]: _type = "Task" [ 769.541765] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 769.552444] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': task-1332118, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 769.780945] env[62753]: DEBUG nova.network.neutron [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 769.861617] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d27f94e8-8298-443f-85fe-a7cdc120cb34 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.865193] env[62753]: DEBUG nova.network.neutron [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 769.871418] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a231b112-1195-4c41-bd04-1da73bc9d204 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.902277] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4da5e81-a418-4ee3-b1e0-b5169e583785 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.910108] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74e99e78-c39c-4bdc-a56d-a58e77cf1bb7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 769.923835] env[62753]: DEBUG nova.compute.provider_tree [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 770.052383] env[62753]: DEBUG oslo_vmware.api [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': task-1332118, 'name': PowerOnVM_Task, 'duration_secs': 0.414049} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 770.052691] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 770.053016] env[62753]: INFO nova.compute.manager [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Took 4.34 seconds to spawn the instance on the hypervisor. [ 770.053300] env[62753]: DEBUG nova.compute.manager [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 770.054078] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6205ea07-3f13-425c-8298-807175911ef3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.095702] env[62753]: DEBUG nova.compute.manager [req-3e138d79-ebbe-48c7-81e2-d4e35f46cb34 req-974d29ad-d290-4115-8b99-956caa753423 service nova] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Received event network-vif-deleted-3df0753e-2da2-4367-9024-eac571d2ebd3 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 770.367900] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Releasing lock "refresh_cache-73f7840f-6e3e-4a4e-b7f5-bffe3009aae6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 770.368355] env[62753]: DEBUG nova.compute.manager [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 770.368549] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 770.368845] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-28703726-fe90-4d08-a6fd-617d6a6a80cd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.377684] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53d5f81e-9d2b-4e32-8fd9-27e72f0961ea {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.401038] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6 could not be found. [ 770.401265] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 770.401441] env[62753]: INFO nova.compute.manager [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Took 0.03 seconds to destroy the instance on the hypervisor. [ 770.401727] env[62753]: DEBUG oslo.service.loopingcall [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 770.401963] env[62753]: DEBUG nova.compute.manager [-] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 770.402072] env[62753]: DEBUG nova.network.neutron [-] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 770.421603] env[62753]: DEBUG nova.network.neutron [-] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 770.427074] env[62753]: DEBUG nova.scheduler.client.report [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 770.508721] env[62753]: DEBUG nova.compute.manager [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 770.532877] env[62753]: DEBUG nova.virt.hardware [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 770.533142] env[62753]: DEBUG nova.virt.hardware [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 770.533295] env[62753]: DEBUG nova.virt.hardware [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 770.533471] env[62753]: DEBUG nova.virt.hardware [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 770.533614] env[62753]: DEBUG nova.virt.hardware [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 770.533814] env[62753]: DEBUG nova.virt.hardware [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 770.534070] env[62753]: DEBUG nova.virt.hardware [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 770.534236] env[62753]: DEBUG nova.virt.hardware [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 770.534406] env[62753]: DEBUG nova.virt.hardware [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 770.534567] env[62753]: DEBUG nova.virt.hardware [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 770.534734] env[62753]: DEBUG nova.virt.hardware [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 770.535600] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90113e4f-e4ce-41ec-9ef9-df659229b2e1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.543781] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06caf618-1837-4171-be3a-97fe1073504b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 770.571876] env[62753]: INFO nova.compute.manager [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Took 22.61 seconds to build instance. [ 770.789585] env[62753]: ERROR nova.compute.manager [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port b2ddb78b-e696-4141-8703-b25192d9afaa, please check neutron logs for more information. [ 770.789585] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 770.789585] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.789585] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 770.789585] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 770.789585] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 770.789585] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 770.789585] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 770.789585] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.789585] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 770.789585] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.789585] env[62753]: ERROR nova.compute.manager raise self.value [ 770.789585] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 770.789585] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 770.789585] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.789585] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 770.790411] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.790411] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 770.790411] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port b2ddb78b-e696-4141-8703-b25192d9afaa, please check neutron logs for more information. [ 770.790411] env[62753]: ERROR nova.compute.manager [ 770.790411] env[62753]: Traceback (most recent call last): [ 770.790411] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 770.790411] env[62753]: listener.cb(fileno) [ 770.790411] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 770.790411] env[62753]: result = function(*args, **kwargs) [ 770.790411] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 770.790411] env[62753]: return func(*args, **kwargs) [ 770.790411] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 770.790411] env[62753]: raise e [ 770.790411] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.790411] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 770.790411] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 770.790411] env[62753]: created_port_ids = self._update_ports_for_instance( [ 770.790411] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 770.790411] env[62753]: with excutils.save_and_reraise_exception(): [ 770.790411] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.790411] env[62753]: self.force_reraise() [ 770.790411] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.790411] env[62753]: raise self.value [ 770.790411] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 770.790411] env[62753]: updated_port = self._update_port( [ 770.790411] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.790411] env[62753]: _ensure_no_port_binding_failure(port) [ 770.790411] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.790411] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 770.791462] env[62753]: nova.exception.PortBindingFailed: Binding failed for port b2ddb78b-e696-4141-8703-b25192d9afaa, please check neutron logs for more information. [ 770.791462] env[62753]: Removing descriptor: 14 [ 770.791462] env[62753]: ERROR nova.compute.manager [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port b2ddb78b-e696-4141-8703-b25192d9afaa, please check neutron logs for more information. [ 770.791462] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Traceback (most recent call last): [ 770.791462] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 770.791462] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] yield resources [ 770.791462] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 770.791462] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] self.driver.spawn(context, instance, image_meta, [ 770.791462] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 770.791462] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 770.791462] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 770.791462] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] vm_ref = self.build_virtual_machine(instance, [ 770.792039] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 770.792039] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] vif_infos = vmwarevif.get_vif_info(self._session, [ 770.792039] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 770.792039] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] for vif in network_info: [ 770.792039] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 770.792039] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] return self._sync_wrapper(fn, *args, **kwargs) [ 770.792039] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 770.792039] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] self.wait() [ 770.792039] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 770.792039] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] self[:] = self._gt.wait() [ 770.792039] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 770.792039] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] return self._exit_event.wait() [ 770.792039] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 770.792466] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] result = hub.switch() [ 770.792466] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 770.792466] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] return self.greenlet.switch() [ 770.792466] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 770.792466] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] result = function(*args, **kwargs) [ 770.792466] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 770.792466] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] return func(*args, **kwargs) [ 770.792466] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 770.792466] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] raise e [ 770.792466] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 770.792466] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] nwinfo = self.network_api.allocate_for_instance( [ 770.792466] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 770.792466] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] created_port_ids = self._update_ports_for_instance( [ 770.792875] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 770.792875] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] with excutils.save_and_reraise_exception(): [ 770.792875] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 770.792875] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] self.force_reraise() [ 770.792875] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 770.792875] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] raise self.value [ 770.792875] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 770.792875] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] updated_port = self._update_port( [ 770.792875] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 770.792875] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] _ensure_no_port_binding_failure(port) [ 770.792875] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 770.792875] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] raise exception.PortBindingFailed(port_id=port['id']) [ 770.793341] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] nova.exception.PortBindingFailed: Binding failed for port b2ddb78b-e696-4141-8703-b25192d9afaa, please check neutron logs for more information. [ 770.793341] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] [ 770.793341] env[62753]: INFO nova.compute.manager [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Terminating instance [ 770.793519] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Acquiring lock "refresh_cache-e260eaf7-98fe-461a-bc69-fc42dfcb79a8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 770.793676] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Acquired lock "refresh_cache-e260eaf7-98fe-461a-bc69-fc42dfcb79a8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 770.793966] env[62753]: DEBUG nova.network.neutron [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 770.923899] env[62753]: DEBUG nova.network.neutron [-] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 770.934360] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.451s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 770.935067] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 770.939170] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.058s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.073951] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fac21fef-a59d-44c8-bdb8-0941dfdb73b4 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Lock "8664ed5c-ecb9-4795-8499-31198cfd0450" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 119.159s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 771.312242] env[62753]: DEBUG nova.network.neutron [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 771.393472] env[62753]: DEBUG nova.network.neutron [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 771.427289] env[62753]: INFO nova.compute.manager [-] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Took 1.03 seconds to deallocate network for instance. [ 771.429411] env[62753]: DEBUG nova.compute.claims [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 771.429559] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.443144] env[62753]: DEBUG nova.compute.utils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 771.446969] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 771.447350] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 771.490052] env[62753]: DEBUG nova.policy [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbfa55fbe33b4312882c8a57a588f6b2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83648a9aa5354de3b0e43d03cea7c1f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 771.578671] env[62753]: DEBUG nova.compute.manager [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 771.722058] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquiring lock "4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 771.722058] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Lock "4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 771.785360] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7f08b9f-ef4d-4ded-be68-6438cf466e19 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.793633] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c176ed57-638b-4bea-8ca6-b9a6fa7b2169 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.823102] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d043d3fe-6397-4cc2-9840-b84e3b62cc29 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.830280] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d122952-db49-4ae9-bb3b-9e02345f391c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.843564] env[62753]: DEBUG nova.compute.provider_tree [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 771.896415] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Releasing lock "refresh_cache-e260eaf7-98fe-461a-bc69-fc42dfcb79a8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 771.896740] env[62753]: DEBUG nova.compute.manager [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 771.896972] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 771.897268] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-15a98a45-84af-432b-8561-4258178010e8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.907348] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b958b09-163e-4f66-9326-aed59dbddc53 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 771.929737] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e260eaf7-98fe-461a-bc69-fc42dfcb79a8 could not be found. [ 771.930053] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 771.930169] env[62753]: INFO nova.compute.manager [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Took 0.03 seconds to destroy the instance on the hypervisor. [ 771.930413] env[62753]: DEBUG oslo.service.loopingcall [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 771.930682] env[62753]: DEBUG nova.compute.manager [-] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 771.931018] env[62753]: DEBUG nova.network.neutron [-] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 771.940031] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Successfully created port: 43c09ea0-9e86-43c5-974f-b35a8f45f8b0 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 771.947552] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 771.950598] env[62753]: DEBUG nova.network.neutron [-] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.032553] env[62753]: DEBUG nova.compute.manager [None req-31cf9416-2c6a-4487-805c-2faafed31af8 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 772.033643] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e718bc10-ea29-4062-91c1-3145833c56ca {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 772.097620] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.102218] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Acquiring lock "8664ed5c-ecb9-4795-8499-31198cfd0450" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.102434] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Lock "8664ed5c-ecb9-4795-8499-31198cfd0450" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.102623] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Acquiring lock "8664ed5c-ecb9-4795-8499-31198cfd0450-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.102870] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Lock "8664ed5c-ecb9-4795-8499-31198cfd0450-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.103071] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Lock "8664ed5c-ecb9-4795-8499-31198cfd0450-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.104973] env[62753]: INFO nova.compute.manager [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Terminating instance [ 772.107116] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Acquiring lock "refresh_cache-8664ed5c-ecb9-4795-8499-31198cfd0450" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.107277] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Acquired lock "refresh_cache-8664ed5c-ecb9-4795-8499-31198cfd0450" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.107437] env[62753]: DEBUG nova.network.neutron [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 772.122331] env[62753]: DEBUG nova.compute.manager [req-a7a2bca0-3aac-46f1-9d0b-e4975b42e247 req-5cf91902-c473-480d-9e12-7ec2fbffe841 service nova] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Received event network-changed-b2ddb78b-e696-4141-8703-b25192d9afaa {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 772.122558] env[62753]: DEBUG nova.compute.manager [req-a7a2bca0-3aac-46f1-9d0b-e4975b42e247 req-5cf91902-c473-480d-9e12-7ec2fbffe841 service nova] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Refreshing instance network info cache due to event network-changed-b2ddb78b-e696-4141-8703-b25192d9afaa. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 772.122743] env[62753]: DEBUG oslo_concurrency.lockutils [req-a7a2bca0-3aac-46f1-9d0b-e4975b42e247 req-5cf91902-c473-480d-9e12-7ec2fbffe841 service nova] Acquiring lock "refresh_cache-e260eaf7-98fe-461a-bc69-fc42dfcb79a8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.122823] env[62753]: DEBUG oslo_concurrency.lockutils [req-a7a2bca0-3aac-46f1-9d0b-e4975b42e247 req-5cf91902-c473-480d-9e12-7ec2fbffe841 service nova] Acquired lock "refresh_cache-e260eaf7-98fe-461a-bc69-fc42dfcb79a8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.123375] env[62753]: DEBUG nova.network.neutron [req-a7a2bca0-3aac-46f1-9d0b-e4975b42e247 req-5cf91902-c473-480d-9e12-7ec2fbffe841 service nova] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Refreshing network info cache for port b2ddb78b-e696-4141-8703-b25192d9afaa {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 772.346687] env[62753]: DEBUG nova.scheduler.client.report [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 772.456205] env[62753]: DEBUG nova.network.neutron [-] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.549658] env[62753]: INFO nova.compute.manager [None req-31cf9416-2c6a-4487-805c-2faafed31af8 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] instance snapshotting [ 772.550374] env[62753]: DEBUG nova.objects.instance [None req-31cf9416-2c6a-4487-805c-2faafed31af8 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Lazy-loading 'flavor' on Instance uuid 8664ed5c-ecb9-4795-8499-31198cfd0450 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 772.626242] env[62753]: DEBUG nova.network.neutron [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.654864] env[62753]: DEBUG nova.network.neutron [req-a7a2bca0-3aac-46f1-9d0b-e4975b42e247 req-5cf91902-c473-480d-9e12-7ec2fbffe841 service nova] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 772.719153] env[62753]: DEBUG nova.network.neutron [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.758225] env[62753]: DEBUG nova.network.neutron [req-a7a2bca0-3aac-46f1-9d0b-e4975b42e247 req-5cf91902-c473-480d-9e12-7ec2fbffe841 service nova] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 772.857032] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.913s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 772.857032] env[62753]: ERROR nova.compute.manager [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 09388d08-50e1-48bc-a122-f3442c3ec2ca, please check neutron logs for more information. [ 772.857032] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Traceback (most recent call last): [ 772.857032] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 772.857032] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] self.driver.spawn(context, instance, image_meta, [ 772.857032] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 772.857032] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] self._vmops.spawn(context, instance, image_meta, injected_files, [ 772.857032] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 772.857032] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] vm_ref = self.build_virtual_machine(instance, [ 772.857442] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 772.857442] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] vif_infos = vmwarevif.get_vif_info(self._session, [ 772.857442] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 772.857442] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] for vif in network_info: [ 772.857442] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 772.857442] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] return self._sync_wrapper(fn, *args, **kwargs) [ 772.857442] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 772.857442] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] self.wait() [ 772.857442] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 772.857442] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] self[:] = self._gt.wait() [ 772.857442] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 772.857442] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] return self._exit_event.wait() [ 772.857442] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 772.857790] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] result = hub.switch() [ 772.857790] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 772.857790] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] return self.greenlet.switch() [ 772.857790] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 772.857790] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] result = function(*args, **kwargs) [ 772.857790] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 772.857790] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] return func(*args, **kwargs) [ 772.857790] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 772.857790] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] raise e [ 772.857790] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 772.857790] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] nwinfo = self.network_api.allocate_for_instance( [ 772.857790] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 772.857790] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] created_port_ids = self._update_ports_for_instance( [ 772.858155] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 772.858155] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] with excutils.save_and_reraise_exception(): [ 772.858155] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 772.858155] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] self.force_reraise() [ 772.858155] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 772.858155] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] raise self.value [ 772.858155] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 772.858155] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] updated_port = self._update_port( [ 772.858155] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 772.858155] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] _ensure_no_port_binding_failure(port) [ 772.858155] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 772.858155] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] raise exception.PortBindingFailed(port_id=port['id']) [ 772.858473] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] nova.exception.PortBindingFailed: Binding failed for port 09388d08-50e1-48bc-a122-f3442c3ec2ca, please check neutron logs for more information. [ 772.858473] env[62753]: ERROR nova.compute.manager [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] [ 772.858473] env[62753]: DEBUG nova.compute.utils [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Binding failed for port 09388d08-50e1-48bc-a122-f3442c3ec2ca, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 772.859049] env[62753]: DEBUG nova.compute.manager [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Build of instance e755857a-e0bc-40bf-bcbb-a416d243442b was re-scheduled: Binding failed for port 09388d08-50e1-48bc-a122-f3442c3ec2ca, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 772.859663] env[62753]: DEBUG nova.compute.manager [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 772.862023] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Acquiring lock "refresh_cache-e755857a-e0bc-40bf-bcbb-a416d243442b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 772.862023] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Acquired lock "refresh_cache-e755857a-e0bc-40bf-bcbb-a416d243442b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 772.862023] env[62753]: DEBUG nova.network.neutron [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 772.862023] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.343s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 772.962025] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 772.962554] env[62753]: INFO nova.compute.manager [-] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Took 1.03 seconds to deallocate network for instance. [ 772.969545] env[62753]: DEBUG nova.compute.claims [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 772.969984] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 772.991749] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 772.992285] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 772.992591] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 772.993131] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 772.993449] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 772.993759] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 772.994154] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 772.994487] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 772.994869] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 772.995261] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 772.997017] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 772.997017] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-713bfaeb-48bf-4614-9a5d-97167d18051f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.005153] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-419287d6-5a58-4419-b3e2-4363e4124021 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.028089] env[62753]: ERROR nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 43c09ea0-9e86-43c5-974f-b35a8f45f8b0, please check neutron logs for more information. [ 773.028089] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 773.028089] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.028089] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 773.028089] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.028089] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 773.028089] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.028089] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 773.028089] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.028089] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 773.028089] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.028089] env[62753]: ERROR nova.compute.manager raise self.value [ 773.028089] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.028089] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 773.028089] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.028089] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 773.028635] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.028635] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 773.028635] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 43c09ea0-9e86-43c5-974f-b35a8f45f8b0, please check neutron logs for more information. [ 773.028635] env[62753]: ERROR nova.compute.manager [ 773.028635] env[62753]: Traceback (most recent call last): [ 773.028635] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 773.028635] env[62753]: listener.cb(fileno) [ 773.028635] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.028635] env[62753]: result = function(*args, **kwargs) [ 773.028635] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 773.028635] env[62753]: return func(*args, **kwargs) [ 773.028635] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.028635] env[62753]: raise e [ 773.028635] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.028635] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 773.028635] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.028635] env[62753]: created_port_ids = self._update_ports_for_instance( [ 773.028635] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.028635] env[62753]: with excutils.save_and_reraise_exception(): [ 773.028635] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.028635] env[62753]: self.force_reraise() [ 773.028635] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.028635] env[62753]: raise self.value [ 773.028635] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.028635] env[62753]: updated_port = self._update_port( [ 773.028635] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.028635] env[62753]: _ensure_no_port_binding_failure(port) [ 773.028635] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.028635] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 773.029546] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 43c09ea0-9e86-43c5-974f-b35a8f45f8b0, please check neutron logs for more information. [ 773.029546] env[62753]: Removing descriptor: 14 [ 773.029546] env[62753]: ERROR nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 43c09ea0-9e86-43c5-974f-b35a8f45f8b0, please check neutron logs for more information. [ 773.029546] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Traceback (most recent call last): [ 773.029546] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 773.029546] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] yield resources [ 773.029546] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 773.029546] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] self.driver.spawn(context, instance, image_meta, [ 773.029546] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 773.029546] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] self._vmops.spawn(context, instance, image_meta, injected_files, [ 773.029546] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 773.029546] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] vm_ref = self.build_virtual_machine(instance, [ 773.030051] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 773.030051] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] vif_infos = vmwarevif.get_vif_info(self._session, [ 773.030051] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 773.030051] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] for vif in network_info: [ 773.030051] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 773.030051] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] return self._sync_wrapper(fn, *args, **kwargs) [ 773.030051] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 773.030051] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] self.wait() [ 773.030051] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 773.030051] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] self[:] = self._gt.wait() [ 773.030051] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 773.030051] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] return self._exit_event.wait() [ 773.030051] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 773.030474] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] result = hub.switch() [ 773.030474] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 773.030474] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] return self.greenlet.switch() [ 773.030474] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 773.030474] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] result = function(*args, **kwargs) [ 773.030474] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 773.030474] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] return func(*args, **kwargs) [ 773.030474] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 773.030474] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] raise e [ 773.030474] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 773.030474] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] nwinfo = self.network_api.allocate_for_instance( [ 773.030474] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 773.030474] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] created_port_ids = self._update_ports_for_instance( [ 773.030902] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 773.030902] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] with excutils.save_and_reraise_exception(): [ 773.030902] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 773.030902] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] self.force_reraise() [ 773.030902] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 773.030902] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] raise self.value [ 773.030902] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 773.030902] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] updated_port = self._update_port( [ 773.030902] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 773.030902] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] _ensure_no_port_binding_failure(port) [ 773.030902] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 773.030902] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] raise exception.PortBindingFailed(port_id=port['id']) [ 773.031281] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] nova.exception.PortBindingFailed: Binding failed for port 43c09ea0-9e86-43c5-974f-b35a8f45f8b0, please check neutron logs for more information. [ 773.031281] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] [ 773.031281] env[62753]: INFO nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Terminating instance [ 773.031424] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "refresh_cache-7831fd59-7a6d-4f6c-9f46-8cb1e2661c17" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 773.031570] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquired lock "refresh_cache-7831fd59-7a6d-4f6c-9f46-8cb1e2661c17" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 773.031815] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 773.057059] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbb16fcf-4e4e-47a7-8edf-dbe9a03280a1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.074091] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-333493f6-44df-4409-89fe-1a0e4c6a8c5c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.222382] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Releasing lock "refresh_cache-8664ed5c-ecb9-4795-8499-31198cfd0450" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.222810] env[62753]: DEBUG nova.compute.manager [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 773.223052] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 773.223945] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bdce857-6d95-4bb6-b67d-a550821f98c4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.231684] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 773.231940] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a3de5402-fcdc-4cc0-a95f-947e5c049115 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.237437] env[62753]: DEBUG oslo_vmware.api [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Waiting for the task: (returnval){ [ 773.237437] env[62753]: value = "task-1332119" [ 773.237437] env[62753]: _type = "Task" [ 773.237437] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.246683] env[62753]: DEBUG oslo_vmware.api [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': task-1332119, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.260562] env[62753]: DEBUG oslo_concurrency.lockutils [req-a7a2bca0-3aac-46f1-9d0b-e4975b42e247 req-5cf91902-c473-480d-9e12-7ec2fbffe841 service nova] Releasing lock "refresh_cache-e260eaf7-98fe-461a-bc69-fc42dfcb79a8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.260562] env[62753]: DEBUG nova.compute.manager [req-a7a2bca0-3aac-46f1-9d0b-e4975b42e247 req-5cf91902-c473-480d-9e12-7ec2fbffe841 service nova] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Received event network-vif-deleted-b2ddb78b-e696-4141-8703-b25192d9afaa {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 773.385117] env[62753]: DEBUG nova.network.neutron [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.458164] env[62753]: DEBUG nova.network.neutron [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.549030] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 773.585058] env[62753]: DEBUG nova.compute.manager [None req-31cf9416-2c6a-4487-805c-2faafed31af8 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Instance disappeared during snapshot {{(pid=62753) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 773.637274] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 773.655545] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7416562-210a-4146-b040-11ba2023d2f3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.663545] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3063785b-5479-43f4-9b1f-ac34be2402e7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.696762] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f13e92fe-08e8-410b-b4ad-9bf6f174f389 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.704069] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6ab10c5-9798-4acb-9576-1d757e35a0da {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.716974] env[62753]: DEBUG nova.compute.provider_tree [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 773.734814] env[62753]: DEBUG nova.compute.manager [None req-31cf9416-2c6a-4487-805c-2faafed31af8 tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Found 0 images (rotation: 2) {{(pid=62753) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 773.747117] env[62753]: DEBUG oslo_vmware.api [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': task-1332119, 'name': PowerOffVM_Task, 'duration_secs': 0.231011} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 773.747903] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 773.748092] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 773.748341] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-451fec83-3406-4219-b748-079383500070 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.771262] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 773.771262] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 773.771765] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Deleting the datastore file [datastore2] 8664ed5c-ecb9-4795-8499-31198cfd0450 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 773.771765] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-41d54c16-5a65-4f01-b1dc-f292748b93a4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 773.778599] env[62753]: DEBUG oslo_vmware.api [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Waiting for the task: (returnval){ [ 773.778599] env[62753]: value = "task-1332121" [ 773.778599] env[62753]: _type = "Task" [ 773.778599] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 773.787958] env[62753]: DEBUG oslo_vmware.api [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': task-1332121, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 773.960657] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Releasing lock "refresh_cache-e755857a-e0bc-40bf-bcbb-a416d243442b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 773.960846] env[62753]: DEBUG nova.compute.manager [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 773.961051] env[62753]: DEBUG nova.compute.manager [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 773.961230] env[62753]: DEBUG nova.network.neutron [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 773.975760] env[62753]: DEBUG nova.network.neutron [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.140240] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Releasing lock "refresh_cache-7831fd59-7a6d-4f6c-9f46-8cb1e2661c17" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 774.140354] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 774.140523] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 774.140803] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f4102e92-0661-4376-b615-45cf7495487b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.145731] env[62753]: DEBUG nova.compute.manager [req-6e818e43-ea55-48db-a1bb-49fd83b279d9 req-f360ea22-d9e5-4703-98a7-51859da8fa14 service nova] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Received event network-changed-43c09ea0-9e86-43c5-974f-b35a8f45f8b0 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 774.145912] env[62753]: DEBUG nova.compute.manager [req-6e818e43-ea55-48db-a1bb-49fd83b279d9 req-f360ea22-d9e5-4703-98a7-51859da8fa14 service nova] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Refreshing instance network info cache due to event network-changed-43c09ea0-9e86-43c5-974f-b35a8f45f8b0. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 774.146186] env[62753]: DEBUG oslo_concurrency.lockutils [req-6e818e43-ea55-48db-a1bb-49fd83b279d9 req-f360ea22-d9e5-4703-98a7-51859da8fa14 service nova] Acquiring lock "refresh_cache-7831fd59-7a6d-4f6c-9f46-8cb1e2661c17" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.146288] env[62753]: DEBUG oslo_concurrency.lockutils [req-6e818e43-ea55-48db-a1bb-49fd83b279d9 req-f360ea22-d9e5-4703-98a7-51859da8fa14 service nova] Acquired lock "refresh_cache-7831fd59-7a6d-4f6c-9f46-8cb1e2661c17" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.146413] env[62753]: DEBUG nova.network.neutron [req-6e818e43-ea55-48db-a1bb-49fd83b279d9 req-f360ea22-d9e5-4703-98a7-51859da8fa14 service nova] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Refreshing network info cache for port 43c09ea0-9e86-43c5-974f-b35a8f45f8b0 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 774.152457] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37ff0ff3-1f24-446f-8e81-d047e557d900 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 774.176531] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17 could not be found. [ 774.176782] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 774.176966] env[62753]: INFO nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Took 0.04 seconds to destroy the instance on the hypervisor. [ 774.177225] env[62753]: DEBUG oslo.service.loopingcall [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 774.177436] env[62753]: DEBUG nova.compute.manager [-] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 774.177527] env[62753]: DEBUG nova.network.neutron [-] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 774.196508] env[62753]: DEBUG nova.network.neutron [-] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.220586] env[62753]: DEBUG nova.scheduler.client.report [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 774.289393] env[62753]: DEBUG oslo_vmware.api [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Task: {'id': task-1332121, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.149791} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 774.289642] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 774.289825] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 774.289993] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 774.290177] env[62753]: INFO nova.compute.manager [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Took 1.07 seconds to destroy the instance on the hypervisor. [ 774.290409] env[62753]: DEBUG oslo.service.loopingcall [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 774.290590] env[62753]: DEBUG nova.compute.manager [-] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 774.290682] env[62753]: DEBUG nova.network.neutron [-] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 774.318993] env[62753]: DEBUG nova.network.neutron [-] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.478121] env[62753]: DEBUG nova.network.neutron [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.666560] env[62753]: DEBUG nova.network.neutron [req-6e818e43-ea55-48db-a1bb-49fd83b279d9 req-f360ea22-d9e5-4703-98a7-51859da8fa14 service nova] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 774.699328] env[62753]: DEBUG nova.network.neutron [-] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.725570] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.864s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 774.726214] env[62753]: ERROR nova.compute.manager [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b067d0d2-dcbe-4852-920e-c9130d215bdc, please check neutron logs for more information. [ 774.726214] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Traceback (most recent call last): [ 774.726214] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 774.726214] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] self.driver.spawn(context, instance, image_meta, [ 774.726214] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 774.726214] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] self._vmops.spawn(context, instance, image_meta, injected_files, [ 774.726214] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 774.726214] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] vm_ref = self.build_virtual_machine(instance, [ 774.726214] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 774.726214] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] vif_infos = vmwarevif.get_vif_info(self._session, [ 774.726214] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 774.726599] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] for vif in network_info: [ 774.726599] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 774.726599] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] return self._sync_wrapper(fn, *args, **kwargs) [ 774.726599] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 774.726599] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] self.wait() [ 774.726599] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 774.726599] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] self[:] = self._gt.wait() [ 774.726599] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 774.726599] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] return self._exit_event.wait() [ 774.726599] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 774.726599] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] result = hub.switch() [ 774.726599] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 774.726599] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] return self.greenlet.switch() [ 774.727016] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 774.727016] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] result = function(*args, **kwargs) [ 774.727016] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 774.727016] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] return func(*args, **kwargs) [ 774.727016] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 774.727016] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] raise e [ 774.727016] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 774.727016] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] nwinfo = self.network_api.allocate_for_instance( [ 774.727016] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 774.727016] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] created_port_ids = self._update_ports_for_instance( [ 774.727016] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 774.727016] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] with excutils.save_and_reraise_exception(): [ 774.727016] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 774.727433] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] self.force_reraise() [ 774.727433] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 774.727433] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] raise self.value [ 774.727433] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 774.727433] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] updated_port = self._update_port( [ 774.727433] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 774.727433] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] _ensure_no_port_binding_failure(port) [ 774.727433] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 774.727433] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] raise exception.PortBindingFailed(port_id=port['id']) [ 774.727433] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] nova.exception.PortBindingFailed: Binding failed for port b067d0d2-dcbe-4852-920e-c9130d215bdc, please check neutron logs for more information. [ 774.727433] env[62753]: ERROR nova.compute.manager [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] [ 774.727785] env[62753]: DEBUG nova.compute.utils [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Binding failed for port b067d0d2-dcbe-4852-920e-c9130d215bdc, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 774.727940] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.378s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 774.728131] env[62753]: DEBUG nova.objects.instance [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62753) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 774.731097] env[62753]: DEBUG nova.compute.manager [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Build of instance eca070f0-dda0-4c51-b2d3-1251c517804a was re-scheduled: Binding failed for port b067d0d2-dcbe-4852-920e-c9130d215bdc, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 774.731514] env[62753]: DEBUG nova.compute.manager [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 774.731751] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Acquiring lock "refresh_cache-eca070f0-dda0-4c51-b2d3-1251c517804a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 774.731907] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Acquired lock "refresh_cache-eca070f0-dda0-4c51-b2d3-1251c517804a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 774.732079] env[62753]: DEBUG nova.network.neutron [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 774.764687] env[62753]: DEBUG nova.network.neutron [req-6e818e43-ea55-48db-a1bb-49fd83b279d9 req-f360ea22-d9e5-4703-98a7-51859da8fa14 service nova] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.821782] env[62753]: DEBUG nova.network.neutron [-] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 774.981030] env[62753]: INFO nova.compute.manager [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] [instance: e755857a-e0bc-40bf-bcbb-a416d243442b] Took 1.02 seconds to deallocate network for instance. [ 775.204880] env[62753]: INFO nova.compute.manager [-] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Took 1.03 seconds to deallocate network for instance. [ 775.207166] env[62753]: DEBUG nova.compute.claims [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 775.207342] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.251933] env[62753]: DEBUG nova.network.neutron [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 775.266787] env[62753]: DEBUG oslo_concurrency.lockutils [req-6e818e43-ea55-48db-a1bb-49fd83b279d9 req-f360ea22-d9e5-4703-98a7-51859da8fa14 service nova] Releasing lock "refresh_cache-7831fd59-7a6d-4f6c-9f46-8cb1e2661c17" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.267052] env[62753]: DEBUG nova.compute.manager [req-6e818e43-ea55-48db-a1bb-49fd83b279d9 req-f360ea22-d9e5-4703-98a7-51859da8fa14 service nova] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Received event network-vif-deleted-43c09ea0-9e86-43c5-974f-b35a8f45f8b0 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 775.324551] env[62753]: INFO nova.compute.manager [-] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Took 1.03 seconds to deallocate network for instance. [ 775.344370] env[62753]: DEBUG nova.network.neutron [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 775.741051] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9140f974-7862-45ff-9b30-5be95b3c03e6 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.012s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 775.742029] env[62753]: DEBUG oslo_concurrency.lockutils [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.938s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 775.830360] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 775.846788] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Releasing lock "refresh_cache-eca070f0-dda0-4c51-b2d3-1251c517804a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 775.846788] env[62753]: DEBUG nova.compute.manager [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 775.846953] env[62753]: DEBUG nova.compute.manager [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 775.847105] env[62753]: DEBUG nova.network.neutron [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 775.867586] env[62753]: DEBUG nova.network.neutron [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 776.019043] env[62753]: INFO nova.scheduler.client.report [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Deleted allocations for instance e755857a-e0bc-40bf-bcbb-a416d243442b [ 776.371544] env[62753]: DEBUG nova.network.neutron [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 776.509224] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a452ef33-813d-45bf-8b99-e59cf0a8ad5b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.517010] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78f33bce-2948-4723-bc49-cc2102011813 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.547016] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a997236a-716c-4ba5-bdd5-0a3a64e90cdf tempest-ImagesOneServerNegativeTestJSON-14484273 tempest-ImagesOneServerNegativeTestJSON-14484273-project-member] Lock "e755857a-e0bc-40bf-bcbb-a416d243442b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.536s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 776.548922] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b021d7b9-65ef-46d9-a1cd-5646ffd8b51c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.556294] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4b008c0-b27c-4caf-806b-113792f7e236 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 776.569987] env[62753]: DEBUG nova.compute.provider_tree [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 776.873804] env[62753]: INFO nova.compute.manager [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] [instance: eca070f0-dda0-4c51-b2d3-1251c517804a] Took 1.03 seconds to deallocate network for instance. [ 777.055101] env[62753]: DEBUG nova.compute.manager [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 777.073463] env[62753]: DEBUG nova.scheduler.client.report [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 777.577657] env[62753]: DEBUG oslo_concurrency.lockutils [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.836s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 777.578210] env[62753]: ERROR nova.compute.manager [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f53e6ca3-7555-4166-b8d6-3613594ac4f8, please check neutron logs for more information. [ 777.578210] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Traceback (most recent call last): [ 777.578210] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 777.578210] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] self.driver.spawn(context, instance, image_meta, [ 777.578210] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 777.578210] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] self._vmops.spawn(context, instance, image_meta, injected_files, [ 777.578210] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 777.578210] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] vm_ref = self.build_virtual_machine(instance, [ 777.578210] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 777.578210] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] vif_infos = vmwarevif.get_vif_info(self._session, [ 777.578210] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 777.578527] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] for vif in network_info: [ 777.578527] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 777.578527] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] return self._sync_wrapper(fn, *args, **kwargs) [ 777.578527] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 777.578527] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] self.wait() [ 777.578527] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 777.578527] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] self[:] = self._gt.wait() [ 777.578527] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 777.578527] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] return self._exit_event.wait() [ 777.578527] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 777.578527] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] current.throw(*self._exc) [ 777.578527] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 777.578527] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] result = function(*args, **kwargs) [ 777.578968] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 777.578968] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] return func(*args, **kwargs) [ 777.578968] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 777.578968] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] raise e [ 777.578968] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 777.578968] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] nwinfo = self.network_api.allocate_for_instance( [ 777.578968] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 777.578968] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] created_port_ids = self._update_ports_for_instance( [ 777.578968] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 777.578968] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] with excutils.save_and_reraise_exception(): [ 777.578968] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 777.578968] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] self.force_reraise() [ 777.578968] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 777.579386] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] raise self.value [ 777.579386] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 777.579386] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] updated_port = self._update_port( [ 777.579386] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 777.579386] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] _ensure_no_port_binding_failure(port) [ 777.579386] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 777.579386] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] raise exception.PortBindingFailed(port_id=port['id']) [ 777.579386] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] nova.exception.PortBindingFailed: Binding failed for port f53e6ca3-7555-4166-b8d6-3613594ac4f8, please check neutron logs for more information. [ 777.579386] env[62753]: ERROR nova.compute.manager [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] [ 777.579386] env[62753]: DEBUG nova.compute.utils [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Binding failed for port f53e6ca3-7555-4166-b8d6-3613594ac4f8, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 777.580697] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.213s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 777.584752] env[62753]: DEBUG nova.compute.manager [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Build of instance b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4 was re-scheduled: Binding failed for port f53e6ca3-7555-4166-b8d6-3613594ac4f8, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 777.585466] env[62753]: DEBUG nova.compute.manager [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 777.586391] env[62753]: DEBUG oslo_concurrency.lockutils [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Acquiring lock "refresh_cache-b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 777.586391] env[62753]: DEBUG oslo_concurrency.lockutils [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Acquired lock "refresh_cache-b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 777.586391] env[62753]: DEBUG nova.network.neutron [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 777.588991] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 777.911891] env[62753]: INFO nova.scheduler.client.report [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Deleted allocations for instance eca070f0-dda0-4c51-b2d3-1251c517804a [ 778.126462] env[62753]: DEBUG nova.network.neutron [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.287021] env[62753]: DEBUG nova.network.neutron [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 778.421026] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7e08a4ad-5c6c-4d6b-bc42-4bb4531481f3 tempest-ServerRescueTestJSONUnderV235-1366234446 tempest-ServerRescueTestJSONUnderV235-1366234446-project-member] Lock "eca070f0-dda0-4c51-b2d3-1251c517804a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.037s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 778.441393] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc1baac4-df29-45af-a3d9-774c3f834cd2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.449310] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed188ce8-5cd3-4a2f-8e2f-c9f2f27e0009 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.484384] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a10ce2e4-0136-4fff-9d2e-985f44e8e9de {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.492013] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ab4a241-b8fe-489c-9603-222bfa64d434 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 778.505554] env[62753]: DEBUG nova.compute.provider_tree [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 778.791312] env[62753]: DEBUG oslo_concurrency.lockutils [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Releasing lock "refresh_cache-b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 778.791613] env[62753]: DEBUG nova.compute.manager [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 778.792014] env[62753]: DEBUG nova.compute.manager [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 778.792101] env[62753]: DEBUG nova.network.neutron [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 778.816331] env[62753]: DEBUG nova.network.neutron [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 778.923250] env[62753]: DEBUG nova.compute.manager [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 779.008775] env[62753]: DEBUG nova.scheduler.client.report [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 779.318994] env[62753]: DEBUG nova.network.neutron [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 779.445051] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 779.515660] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.935s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 779.516387] env[62753]: ERROR nova.compute.manager [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 0f17649b-6479-4040-b2cf-71fa86dbbe03, please check neutron logs for more information. [ 779.516387] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Traceback (most recent call last): [ 779.516387] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 779.516387] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] self.driver.spawn(context, instance, image_meta, [ 779.516387] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 779.516387] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] self._vmops.spawn(context, instance, image_meta, injected_files, [ 779.516387] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 779.516387] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] vm_ref = self.build_virtual_machine(instance, [ 779.516387] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 779.516387] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] vif_infos = vmwarevif.get_vif_info(self._session, [ 779.516387] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 779.516866] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] for vif in network_info: [ 779.516866] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 779.516866] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] return self._sync_wrapper(fn, *args, **kwargs) [ 779.516866] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 779.516866] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] self.wait() [ 779.516866] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 779.516866] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] self[:] = self._gt.wait() [ 779.516866] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 779.516866] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] return self._exit_event.wait() [ 779.516866] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 779.516866] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] result = hub.switch() [ 779.516866] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 779.516866] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] return self.greenlet.switch() [ 779.517554] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 779.517554] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] result = function(*args, **kwargs) [ 779.517554] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 779.517554] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] return func(*args, **kwargs) [ 779.517554] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 779.517554] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] raise e [ 779.517554] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 779.517554] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] nwinfo = self.network_api.allocate_for_instance( [ 779.517554] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 779.517554] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] created_port_ids = self._update_ports_for_instance( [ 779.517554] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 779.517554] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] with excutils.save_and_reraise_exception(): [ 779.517554] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 779.518184] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] self.force_reraise() [ 779.518184] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 779.518184] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] raise self.value [ 779.518184] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 779.518184] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] updated_port = self._update_port( [ 779.518184] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 779.518184] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] _ensure_no_port_binding_failure(port) [ 779.518184] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 779.518184] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] raise exception.PortBindingFailed(port_id=port['id']) [ 779.518184] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] nova.exception.PortBindingFailed: Binding failed for port 0f17649b-6479-4040-b2cf-71fa86dbbe03, please check neutron logs for more information. [ 779.518184] env[62753]: ERROR nova.compute.manager [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] [ 779.518616] env[62753]: DEBUG nova.compute.utils [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Binding failed for port 0f17649b-6479-4040-b2cf-71fa86dbbe03, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 779.520273] env[62753]: DEBUG nova.compute.manager [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Build of instance 358d1c98-28ee-4f18-bd42-89a1e7687d03 was re-scheduled: Binding failed for port 0f17649b-6479-4040-b2cf-71fa86dbbe03, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 779.522118] env[62753]: DEBUG nova.compute.manager [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 779.522377] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "refresh_cache-358d1c98-28ee-4f18-bd42-89a1e7687d03" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 779.522526] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "refresh_cache-358d1c98-28ee-4f18-bd42-89a1e7687d03" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 779.522685] env[62753]: DEBUG nova.network.neutron [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 779.523719] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.345s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 779.523944] env[62753]: DEBUG nova.objects.instance [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Lazy-loading 'resources' on Instance uuid a8ceac07-6631-4ac6-b62a-7f3eac788f2b {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 779.821474] env[62753]: INFO nova.compute.manager [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] [instance: b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4] Took 1.03 seconds to deallocate network for instance. [ 780.061271] env[62753]: DEBUG nova.network.neutron [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.177878] env[62753]: DEBUG nova.network.neutron [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 780.358088] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53d6f73b-fe91-4773-ae70-46d38a843a49 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.368021] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35a14681-7b72-4172-a94d-d9aebf8722d7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.403459] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1761b59-a655-42bb-af3d-176fb971dc09 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.410540] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-371f78dd-d9a5-48e4-a86f-17a02c4cee09 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 780.424394] env[62753]: DEBUG nova.compute.provider_tree [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 780.680872] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "refresh_cache-358d1c98-28ee-4f18-bd42-89a1e7687d03" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 780.681194] env[62753]: DEBUG nova.compute.manager [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 780.681438] env[62753]: DEBUG nova.compute.manager [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 780.681656] env[62753]: DEBUG nova.network.neutron [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 780.696642] env[62753]: DEBUG nova.network.neutron [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 780.854474] env[62753]: INFO nova.scheduler.client.report [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Deleted allocations for instance b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4 [ 780.930026] env[62753]: DEBUG nova.scheduler.client.report [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 781.200271] env[62753]: DEBUG nova.network.neutron [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 781.363911] env[62753]: DEBUG oslo_concurrency.lockutils [None req-16903126-004d-42c8-af3d-a2e0bed4d168 tempest-ServerAddressesTestJSON-1903238546 tempest-ServerAddressesTestJSON-1903238546-project-member] Lock "b6e236f9-c5fa-40e2-b445-49ca1fd6cbf4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 139.545s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.437022] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.910s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 781.437022] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 17.656s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 781.457062] env[62753]: INFO nova.scheduler.client.report [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Deleted allocations for instance a8ceac07-6631-4ac6-b62a-7f3eac788f2b [ 781.704600] env[62753]: INFO nova.compute.manager [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 358d1c98-28ee-4f18-bd42-89a1e7687d03] Took 1.02 seconds to deallocate network for instance. [ 781.865819] env[62753]: DEBUG nova.compute.manager [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 781.967267] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c7431d72-ba52-41fa-ac52-4be304c49a51 tempest-ServerShowV257Test-1598075492 tempest-ServerShowV257Test-1598075492-project-member] Lock "a8ceac07-6631-4ac6-b62a-7f3eac788f2b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.483s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 782.253878] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b85fc51-61e3-4165-9ed6-80e6757f8746 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.263097] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e296af-d653-4285-aa48-a9c8c42a04f9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.303994] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-158e7a0d-3cbd-4d34-96e4-4aa4c0cd54da {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.311512] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9b345d-a215-4c8d-ae9b-79dcfa9db60b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 782.325452] env[62753]: DEBUG nova.compute.provider_tree [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 782.387786] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 782.743755] env[62753]: INFO nova.scheduler.client.report [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Deleted allocations for instance 358d1c98-28ee-4f18-bd42-89a1e7687d03 [ 782.828829] env[62753]: DEBUG nova.scheduler.client.report [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 783.255137] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ce9ccf6-bb94-4c64-bccb-e36e17a6be38 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "358d1c98-28ee-4f18-bd42-89a1e7687d03" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.791s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.338024] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.899s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 783.338024] env[62753]: ERROR nova.compute.manager [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 7d800493-1971-4372-8c86-b9d4ff0be887, please check neutron logs for more information. [ 783.338024] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Traceback (most recent call last): [ 783.338024] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 783.338024] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] self.driver.spawn(context, instance, image_meta, [ 783.338024] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 783.338024] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] self._vmops.spawn(context, instance, image_meta, injected_files, [ 783.338024] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 783.338024] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] vm_ref = self.build_virtual_machine(instance, [ 783.338380] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 783.338380] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] vif_infos = vmwarevif.get_vif_info(self._session, [ 783.338380] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 783.338380] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] for vif in network_info: [ 783.338380] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 783.338380] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] return self._sync_wrapper(fn, *args, **kwargs) [ 783.338380] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 783.338380] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] self.wait() [ 783.338380] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 783.338380] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] self[:] = self._gt.wait() [ 783.338380] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 783.338380] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] return self._exit_event.wait() [ 783.338380] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 783.338725] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] result = hub.switch() [ 783.338725] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 783.338725] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] return self.greenlet.switch() [ 783.338725] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 783.338725] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] result = function(*args, **kwargs) [ 783.338725] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 783.338725] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] return func(*args, **kwargs) [ 783.338725] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 783.338725] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] raise e [ 783.338725] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 783.338725] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] nwinfo = self.network_api.allocate_for_instance( [ 783.338725] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 783.338725] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] created_port_ids = self._update_ports_for_instance( [ 783.339092] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 783.339092] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] with excutils.save_and_reraise_exception(): [ 783.339092] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 783.339092] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] self.force_reraise() [ 783.339092] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 783.339092] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] raise self.value [ 783.339092] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 783.339092] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] updated_port = self._update_port( [ 783.339092] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 783.339092] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] _ensure_no_port_binding_failure(port) [ 783.339092] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 783.339092] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] raise exception.PortBindingFailed(port_id=port['id']) [ 783.339412] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] nova.exception.PortBindingFailed: Binding failed for port 7d800493-1971-4372-8c86-b9d4ff0be887, please check neutron logs for more information. [ 783.339412] env[62753]: ERROR nova.compute.manager [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] [ 783.339412] env[62753]: DEBUG nova.compute.utils [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Binding failed for port 7d800493-1971-4372-8c86-b9d4ff0be887, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 783.347046] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.274s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 783.347046] env[62753]: INFO nova.compute.claims [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 783.348800] env[62753]: DEBUG nova.compute.manager [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Build of instance 365fe802-f5e1-47d6-807e-a3b1027def02 was re-scheduled: Binding failed for port 7d800493-1971-4372-8c86-b9d4ff0be887, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 783.349666] env[62753]: DEBUG nova.compute.manager [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 783.350026] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "refresh_cache-365fe802-f5e1-47d6-807e-a3b1027def02" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 783.352165] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquired lock "refresh_cache-365fe802-f5e1-47d6-807e-a3b1027def02" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 783.352547] env[62753]: DEBUG nova.network.neutron [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 783.759218] env[62753]: DEBUG nova.compute.manager [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 783.880644] env[62753]: DEBUG nova.network.neutron [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.013073] env[62753]: DEBUG nova.network.neutron [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 784.287694] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 784.516525] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Releasing lock "refresh_cache-365fe802-f5e1-47d6-807e-a3b1027def02" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 784.516825] env[62753]: DEBUG nova.compute.manager [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 784.517027] env[62753]: DEBUG nova.compute.manager [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 784.517206] env[62753]: DEBUG nova.network.neutron [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 784.560987] env[62753]: DEBUG nova.network.neutron [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 784.657688] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a3c039d-e7d5-4e7c-abd5-d0917ade8134 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.665868] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32fd9788-9968-4418-88b5-ab9f72d2b50b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.700030] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ba92bbc-a0ae-4b2d-89b8-28948b215df6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.708440] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fd8de84-1d24-4da5-a3ad-37f880384207 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 784.722331] env[62753]: DEBUG nova.compute.provider_tree [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 785.026741] env[62753]: DEBUG oslo_concurrency.lockutils [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Acquiring lock "f132230d-0fba-4293-b940-85cc7ef1cad1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.027043] env[62753]: DEBUG oslo_concurrency.lockutils [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Lock "f132230d-0fba-4293-b940-85cc7ef1cad1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.066590] env[62753]: DEBUG nova.network.neutron [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 785.228888] env[62753]: DEBUG nova.scheduler.client.report [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 785.569478] env[62753]: INFO nova.compute.manager [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 365fe802-f5e1-47d6-807e-a3b1027def02] Took 1.05 seconds to deallocate network for instance. [ 785.583417] env[62753]: DEBUG oslo_concurrency.lockutils [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "a6ee010c-6305-4009-80e0-92a2c58bcd7b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 785.583644] env[62753]: DEBUG oslo_concurrency.lockutils [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "a6ee010c-6305-4009-80e0-92a2c58bcd7b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.733337] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.389s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.733909] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 785.736759] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.848s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.736949] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 785.737159] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62753) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 785.738031] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 14.308s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 785.742193] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b39279ae-230d-41be-81eb-0394162a94ef {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.753353] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b30d7e1-7e1f-417e-bcad-c281c45c69f2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.770764] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aee6509-4207-4ec8-a912-56fb0f0b3b97 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.781534] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbd38716-2327-483c-8435-ba8de08d2547 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 785.827158] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181456MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=62753) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 785.827312] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 786.245768] env[62753]: DEBUG nova.compute.utils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 786.247498] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 786.247684] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 786.296946] env[62753]: DEBUG nova.policy [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'fbfa55fbe33b4312882c8a57a588f6b2', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '83648a9aa5354de3b0e43d03cea7c1f0', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 786.492447] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a2e1b2c-1903-4f79-a9ba-d9483b0c8633 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.500123] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d9890c2-1228-4db8-bd29-94769ad31a3f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.533718] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ddce4b3-e651-4dfb-9ee4-01d937834ed0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.542116] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8acec56d-618a-4218-ac64-50fae5f71f01 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 786.557301] env[62753]: DEBUG nova.compute.provider_tree [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 786.604788] env[62753]: INFO nova.scheduler.client.report [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Deleted allocations for instance 365fe802-f5e1-47d6-807e-a3b1027def02 [ 786.630047] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Successfully created port: 53d194f5-a7ce-4eb0-99da-e89986b6fc73 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 786.750905] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 787.060533] env[62753]: DEBUG nova.scheduler.client.report [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 787.120093] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62d3ec2b-9f62-4deb-b364-59bd08bc4721 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "365fe802-f5e1-47d6-807e-a3b1027def02" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 140.255s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.496160] env[62753]: DEBUG nova.compute.manager [req-1e7720a3-e35b-4f12-ac93-fc622f491c01 req-b777f607-c937-4c13-9736-6c81e17a6941 service nova] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Received event network-changed-53d194f5-a7ce-4eb0-99da-e89986b6fc73 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 787.496160] env[62753]: DEBUG nova.compute.manager [req-1e7720a3-e35b-4f12-ac93-fc622f491c01 req-b777f607-c937-4c13-9736-6c81e17a6941 service nova] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Refreshing instance network info cache due to event network-changed-53d194f5-a7ce-4eb0-99da-e89986b6fc73. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 787.496160] env[62753]: DEBUG oslo_concurrency.lockutils [req-1e7720a3-e35b-4f12-ac93-fc622f491c01 req-b777f607-c937-4c13-9736-6c81e17a6941 service nova] Acquiring lock "refresh_cache-d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.496160] env[62753]: DEBUG oslo_concurrency.lockutils [req-1e7720a3-e35b-4f12-ac93-fc622f491c01 req-b777f607-c937-4c13-9736-6c81e17a6941 service nova] Acquired lock "refresh_cache-d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.496160] env[62753]: DEBUG nova.network.neutron [req-1e7720a3-e35b-4f12-ac93-fc622f491c01 req-b777f607-c937-4c13-9736-6c81e17a6941 service nova] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Refreshing network info cache for port 53d194f5-a7ce-4eb0-99da-e89986b6fc73 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 787.570643] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.831s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 787.570643] env[62753]: ERROR nova.compute.manager [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 3df0753e-2da2-4367-9024-eac571d2ebd3, please check neutron logs for more information. [ 787.570643] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Traceback (most recent call last): [ 787.570643] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 787.570643] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] self.driver.spawn(context, instance, image_meta, [ 787.570643] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 787.570643] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] self._vmops.spawn(context, instance, image_meta, injected_files, [ 787.570643] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 787.570643] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] vm_ref = self.build_virtual_machine(instance, [ 787.571040] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 787.571040] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] vif_infos = vmwarevif.get_vif_info(self._session, [ 787.571040] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 787.571040] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] for vif in network_info: [ 787.571040] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 787.571040] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] return self._sync_wrapper(fn, *args, **kwargs) [ 787.571040] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 787.571040] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] self.wait() [ 787.571040] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 787.571040] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] self[:] = self._gt.wait() [ 787.571040] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 787.571040] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] return self._exit_event.wait() [ 787.571040] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 787.571408] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] result = hub.switch() [ 787.571408] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 787.571408] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] return self.greenlet.switch() [ 787.571408] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 787.571408] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] result = function(*args, **kwargs) [ 787.571408] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 787.571408] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] return func(*args, **kwargs) [ 787.571408] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 787.571408] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] raise e [ 787.571408] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 787.571408] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] nwinfo = self.network_api.allocate_for_instance( [ 787.571408] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.571408] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] created_port_ids = self._update_ports_for_instance( [ 787.571748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.571748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] with excutils.save_and_reraise_exception(): [ 787.571748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.571748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] self.force_reraise() [ 787.571748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.571748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] raise self.value [ 787.571748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.571748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] updated_port = self._update_port( [ 787.571748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.571748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] _ensure_no_port_binding_failure(port) [ 787.571748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.571748] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] raise exception.PortBindingFailed(port_id=port['id']) [ 787.572106] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] nova.exception.PortBindingFailed: Binding failed for port 3df0753e-2da2-4367-9024-eac571d2ebd3, please check neutron logs for more information. [ 787.572106] env[62753]: ERROR nova.compute.manager [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] [ 787.572106] env[62753]: DEBUG nova.compute.utils [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Binding failed for port 3df0753e-2da2-4367-9024-eac571d2ebd3, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 787.572106] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.474s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 787.573129] env[62753]: INFO nova.compute.claims [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 787.576095] env[62753]: DEBUG nova.compute.manager [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Build of instance 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6 was re-scheduled: Binding failed for port 3df0753e-2da2-4367-9024-eac571d2ebd3, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 787.576542] env[62753]: DEBUG nova.compute.manager [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 787.576803] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Acquiring lock "refresh_cache-73f7840f-6e3e-4a4e-b7f5-bffe3009aae6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 787.577083] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Acquired lock "refresh_cache-73f7840f-6e3e-4a4e-b7f5-bffe3009aae6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 787.577212] env[62753]: DEBUG nova.network.neutron [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 787.623362] env[62753]: DEBUG nova.compute.manager [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 787.701705] env[62753]: ERROR nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 53d194f5-a7ce-4eb0-99da-e89986b6fc73, please check neutron logs for more information. [ 787.701705] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 787.701705] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 787.701705] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 787.701705] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.701705] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 787.701705] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.701705] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 787.701705] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.701705] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 787.701705] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.701705] env[62753]: ERROR nova.compute.manager raise self.value [ 787.701705] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.701705] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 787.701705] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.701705] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 787.702244] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.702244] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 787.702244] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 53d194f5-a7ce-4eb0-99da-e89986b6fc73, please check neutron logs for more information. [ 787.702244] env[62753]: ERROR nova.compute.manager [ 787.702244] env[62753]: Traceback (most recent call last): [ 787.702244] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 787.702244] env[62753]: listener.cb(fileno) [ 787.702244] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 787.702244] env[62753]: result = function(*args, **kwargs) [ 787.702244] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 787.702244] env[62753]: return func(*args, **kwargs) [ 787.702244] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 787.702244] env[62753]: raise e [ 787.702244] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 787.702244] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 787.702244] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.702244] env[62753]: created_port_ids = self._update_ports_for_instance( [ 787.702244] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.702244] env[62753]: with excutils.save_and_reraise_exception(): [ 787.702244] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.702244] env[62753]: self.force_reraise() [ 787.702244] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.702244] env[62753]: raise self.value [ 787.702244] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.702244] env[62753]: updated_port = self._update_port( [ 787.702244] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.702244] env[62753]: _ensure_no_port_binding_failure(port) [ 787.702244] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.702244] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 787.703082] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 53d194f5-a7ce-4eb0-99da-e89986b6fc73, please check neutron logs for more information. [ 787.703082] env[62753]: Removing descriptor: 14 [ 787.763686] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 787.797530] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 787.797831] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 787.798015] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 787.798809] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 787.798809] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 787.798809] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 787.798809] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 787.798809] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 787.799064] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 787.799131] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 787.799309] env[62753]: DEBUG nova.virt.hardware [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 787.800606] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7bf294d-2462-475b-aa57-38817a3484da {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.808738] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-638d571f-70df-4cbf-a459-67d252e93106 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 787.822680] env[62753]: ERROR nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 53d194f5-a7ce-4eb0-99da-e89986b6fc73, please check neutron logs for more information. [ 787.822680] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Traceback (most recent call last): [ 787.822680] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 787.822680] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] yield resources [ 787.822680] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 787.822680] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] self.driver.spawn(context, instance, image_meta, [ 787.822680] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 787.822680] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] self._vmops.spawn(context, instance, image_meta, injected_files, [ 787.822680] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 787.822680] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] vm_ref = self.build_virtual_machine(instance, [ 787.822680] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 787.823152] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] vif_infos = vmwarevif.get_vif_info(self._session, [ 787.823152] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 787.823152] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] for vif in network_info: [ 787.823152] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 787.823152] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] return self._sync_wrapper(fn, *args, **kwargs) [ 787.823152] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 787.823152] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] self.wait() [ 787.823152] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 787.823152] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] self[:] = self._gt.wait() [ 787.823152] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 787.823152] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] return self._exit_event.wait() [ 787.823152] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 787.823152] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] current.throw(*self._exc) [ 787.823527] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 787.823527] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] result = function(*args, **kwargs) [ 787.823527] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 787.823527] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] return func(*args, **kwargs) [ 787.823527] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 787.823527] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] raise e [ 787.823527] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 787.823527] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] nwinfo = self.network_api.allocate_for_instance( [ 787.823527] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 787.823527] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] created_port_ids = self._update_ports_for_instance( [ 787.823527] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 787.823527] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] with excutils.save_and_reraise_exception(): [ 787.823527] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 787.823905] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] self.force_reraise() [ 787.823905] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 787.823905] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] raise self.value [ 787.823905] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 787.823905] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] updated_port = self._update_port( [ 787.823905] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 787.823905] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] _ensure_no_port_binding_failure(port) [ 787.823905] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 787.823905] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] raise exception.PortBindingFailed(port_id=port['id']) [ 787.823905] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] nova.exception.PortBindingFailed: Binding failed for port 53d194f5-a7ce-4eb0-99da-e89986b6fc73, please check neutron logs for more information. [ 787.823905] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] [ 787.823905] env[62753]: INFO nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Terminating instance [ 787.825068] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "refresh_cache-d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 788.020918] env[62753]: DEBUG nova.network.neutron [req-1e7720a3-e35b-4f12-ac93-fc622f491c01 req-b777f607-c937-4c13-9736-6c81e17a6941 service nova] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.102686] env[62753]: DEBUG nova.network.neutron [req-1e7720a3-e35b-4f12-ac93-fc622f491c01 req-b777f607-c937-4c13-9736-6c81e17a6941 service nova] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.104377] env[62753]: DEBUG nova.network.neutron [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.156116] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 788.230314] env[62753]: DEBUG nova.network.neutron [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 788.607729] env[62753]: DEBUG oslo_concurrency.lockutils [req-1e7720a3-e35b-4f12-ac93-fc622f491c01 req-b777f607-c937-4c13-9736-6c81e17a6941 service nova] Releasing lock "refresh_cache-d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.608197] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquired lock "refresh_cache-d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 788.608393] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 788.732458] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Releasing lock "refresh_cache-73f7840f-6e3e-4a4e-b7f5-bffe3009aae6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 788.732691] env[62753]: DEBUG nova.compute.manager [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 788.732864] env[62753]: DEBUG nova.compute.manager [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 788.733671] env[62753]: DEBUG nova.network.neutron [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 788.757986] env[62753]: DEBUG nova.network.neutron [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 788.836843] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30a00bd1-ef4a-45fa-9b0e-a8465f2590b9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.845701] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab159139-590f-4eaf-8143-ad45987e8ed9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.878478] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f211e40d-cc6f-4569-956e-c6e1b461b899 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.885523] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-180b6efb-1922-4b3e-a88f-5e10d303b3cd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 788.898738] env[62753]: DEBUG nova.compute.provider_tree [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 789.127644] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.203098] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.261006] env[62753]: DEBUG nova.network.neutron [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 789.402929] env[62753]: DEBUG nova.scheduler.client.report [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 789.522523] env[62753]: DEBUG nova.compute.manager [req-f450ebb6-4e67-4093-b867-5f72c413fcdf req-a4a67925-05aa-4537-ae75-d2e3fa32e581 service nova] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Received event network-vif-deleted-53d194f5-a7ce-4eb0-99da-e89986b6fc73 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 789.708689] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Releasing lock "refresh_cache-d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 789.709115] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 789.709313] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 789.709604] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5ff6d84c-948b-4e3d-9390-fa32a16aff65 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.719100] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c426b33a-ecdd-4baf-8520-ba45e3197bfc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 789.740230] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295 could not be found. [ 789.740317] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 789.740535] env[62753]: INFO nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Took 0.03 seconds to destroy the instance on the hypervisor. [ 789.740737] env[62753]: DEBUG oslo.service.loopingcall [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 789.740961] env[62753]: DEBUG nova.compute.manager [-] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 789.741069] env[62753]: DEBUG nova.network.neutron [-] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 789.757274] env[62753]: DEBUG nova.network.neutron [-] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 789.763508] env[62753]: INFO nova.compute.manager [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] [instance: 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6] Took 1.03 seconds to deallocate network for instance. [ 789.909715] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.338s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 789.910292] env[62753]: DEBUG nova.compute.manager [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 789.912913] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.943s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.035233] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.035490] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 790.259737] env[62753]: DEBUG nova.network.neutron [-] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 790.419959] env[62753]: DEBUG nova.compute.utils [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 790.421474] env[62753]: DEBUG nova.compute.manager [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 790.421620] env[62753]: DEBUG nova.network.neutron [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 790.473185] env[62753]: DEBUG nova.policy [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '13dbab5a694f4b7fa3bc55a9fe47d8ba', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8a3ac082259b4014a49c9df5097f9e9d', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 790.655638] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-967c0caf-962f-4886-acda-aecf3cbda1c6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.663435] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75af1c5d-30fa-4227-8c65-d35b2e17c2bc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.692891] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09191e66-4f1c-4f76-be51-1e7ac15bd8ed {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.700368] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0de33821-f73c-4f8d-ae87-721a28400d85 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 790.713064] env[62753]: DEBUG nova.compute.provider_tree [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 790.725710] env[62753]: DEBUG nova.network.neutron [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Successfully created port: 257c7620-1533-42cc-b45b-b6615400c2d3 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 790.761952] env[62753]: INFO nova.compute.manager [-] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Took 1.02 seconds to deallocate network for instance. [ 790.764546] env[62753]: DEBUG nova.compute.claims [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 790.764732] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 790.791627] env[62753]: INFO nova.scheduler.client.report [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Deleted allocations for instance 73f7840f-6e3e-4a4e-b7f5-bffe3009aae6 [ 790.926673] env[62753]: DEBUG nova.compute.manager [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 791.218535] env[62753]: DEBUG nova.scheduler.client.report [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 791.304098] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d4ecc7a3-0757-4a5d-b0cd-6ebe4796e154 tempest-ServersTestManualDisk-1420849830 tempest-ServersTestManualDisk-1420849830-project-member] Lock "73f7840f-6e3e-4a4e-b7f5-bffe3009aae6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 138.678s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.381866] env[62753]: DEBUG nova.compute.manager [req-376e911e-a1af-46f2-a81f-77f672f39a83 req-7d50b731-f313-48bb-a4b2-f31bdf121f15 service nova] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Received event network-changed-257c7620-1533-42cc-b45b-b6615400c2d3 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 791.382177] env[62753]: DEBUG nova.compute.manager [req-376e911e-a1af-46f2-a81f-77f672f39a83 req-7d50b731-f313-48bb-a4b2-f31bdf121f15 service nova] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Refreshing instance network info cache due to event network-changed-257c7620-1533-42cc-b45b-b6615400c2d3. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 791.382305] env[62753]: DEBUG oslo_concurrency.lockutils [req-376e911e-a1af-46f2-a81f-77f672f39a83 req-7d50b731-f313-48bb-a4b2-f31bdf121f15 service nova] Acquiring lock "refresh_cache-e0c4a1d2-6a85-4826-88fa-d21556d744ac" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.382400] env[62753]: DEBUG oslo_concurrency.lockutils [req-376e911e-a1af-46f2-a81f-77f672f39a83 req-7d50b731-f313-48bb-a4b2-f31bdf121f15 service nova] Acquired lock "refresh_cache-e0c4a1d2-6a85-4826-88fa-d21556d744ac" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.382537] env[62753]: DEBUG nova.network.neutron [req-376e911e-a1af-46f2-a81f-77f672f39a83 req-7d50b731-f313-48bb-a4b2-f31bdf121f15 service nova] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Refreshing network info cache for port 257c7620-1533-42cc-b45b-b6615400c2d3 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 791.553507] env[62753]: ERROR nova.compute.manager [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 257c7620-1533-42cc-b45b-b6615400c2d3, please check neutron logs for more information. [ 791.553507] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 791.553507] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.553507] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 791.553507] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.553507] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 791.553507] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.553507] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 791.553507] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.553507] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 791.553507] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.553507] env[62753]: ERROR nova.compute.manager raise self.value [ 791.553507] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.553507] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 791.553507] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.553507] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 791.554252] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.554252] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 791.554252] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 257c7620-1533-42cc-b45b-b6615400c2d3, please check neutron logs for more information. [ 791.554252] env[62753]: ERROR nova.compute.manager [ 791.554252] env[62753]: Traceback (most recent call last): [ 791.554252] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 791.554252] env[62753]: listener.cb(fileno) [ 791.554252] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 791.554252] env[62753]: result = function(*args, **kwargs) [ 791.554252] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 791.554252] env[62753]: return func(*args, **kwargs) [ 791.554252] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 791.554252] env[62753]: raise e [ 791.554252] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.554252] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 791.554252] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.554252] env[62753]: created_port_ids = self._update_ports_for_instance( [ 791.554252] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.554252] env[62753]: with excutils.save_and_reraise_exception(): [ 791.554252] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.554252] env[62753]: self.force_reraise() [ 791.554252] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.554252] env[62753]: raise self.value [ 791.554252] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.554252] env[62753]: updated_port = self._update_port( [ 791.554252] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.554252] env[62753]: _ensure_no_port_binding_failure(port) [ 791.554252] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.554252] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 791.555384] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 257c7620-1533-42cc-b45b-b6615400c2d3, please check neutron logs for more information. [ 791.555384] env[62753]: Removing descriptor: 18 [ 791.722722] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.810s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 791.723411] env[62753]: ERROR nova.compute.manager [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port b2ddb78b-e696-4141-8703-b25192d9afaa, please check neutron logs for more information. [ 791.723411] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Traceback (most recent call last): [ 791.723411] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 791.723411] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] self.driver.spawn(context, instance, image_meta, [ 791.723411] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 791.723411] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] self._vmops.spawn(context, instance, image_meta, injected_files, [ 791.723411] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 791.723411] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] vm_ref = self.build_virtual_machine(instance, [ 791.723411] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 791.723411] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] vif_infos = vmwarevif.get_vif_info(self._session, [ 791.723411] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 791.723885] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] for vif in network_info: [ 791.723885] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 791.723885] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] return self._sync_wrapper(fn, *args, **kwargs) [ 791.723885] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 791.723885] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] self.wait() [ 791.723885] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 791.723885] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] self[:] = self._gt.wait() [ 791.723885] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 791.723885] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] return self._exit_event.wait() [ 791.723885] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 791.723885] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] result = hub.switch() [ 791.723885] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 791.723885] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] return self.greenlet.switch() [ 791.724286] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 791.724286] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] result = function(*args, **kwargs) [ 791.724286] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 791.724286] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] return func(*args, **kwargs) [ 791.724286] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 791.724286] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] raise e [ 791.724286] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.724286] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] nwinfo = self.network_api.allocate_for_instance( [ 791.724286] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.724286] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] created_port_ids = self._update_ports_for_instance( [ 791.724286] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.724286] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] with excutils.save_and_reraise_exception(): [ 791.724286] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.724632] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] self.force_reraise() [ 791.724632] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.724632] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] raise self.value [ 791.724632] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.724632] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] updated_port = self._update_port( [ 791.724632] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.724632] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] _ensure_no_port_binding_failure(port) [ 791.724632] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.724632] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] raise exception.PortBindingFailed(port_id=port['id']) [ 791.724632] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] nova.exception.PortBindingFailed: Binding failed for port b2ddb78b-e696-4141-8703-b25192d9afaa, please check neutron logs for more information. [ 791.724632] env[62753]: ERROR nova.compute.manager [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] [ 791.724932] env[62753]: DEBUG nova.compute.utils [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Binding failed for port b2ddb78b-e696-4141-8703-b25192d9afaa, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 791.725279] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.518s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 791.728758] env[62753]: DEBUG nova.compute.manager [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Build of instance e260eaf7-98fe-461a-bc69-fc42dfcb79a8 was re-scheduled: Binding failed for port b2ddb78b-e696-4141-8703-b25192d9afaa, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 791.729269] env[62753]: DEBUG nova.compute.manager [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 791.729519] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Acquiring lock "refresh_cache-e260eaf7-98fe-461a-bc69-fc42dfcb79a8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 791.729668] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Acquired lock "refresh_cache-e260eaf7-98fe-461a-bc69-fc42dfcb79a8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 791.730804] env[62753]: DEBUG nova.network.neutron [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 791.804739] env[62753]: DEBUG nova.compute.manager [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 791.905122] env[62753]: DEBUG nova.network.neutron [req-376e911e-a1af-46f2-a81f-77f672f39a83 req-7d50b731-f313-48bb-a4b2-f31bdf121f15 service nova] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 791.935838] env[62753]: DEBUG nova.compute.manager [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 791.960013] env[62753]: DEBUG nova.virt.hardware [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 791.960280] env[62753]: DEBUG nova.virt.hardware [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 791.960435] env[62753]: DEBUG nova.virt.hardware [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 791.960782] env[62753]: DEBUG nova.virt.hardware [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 791.960859] env[62753]: DEBUG nova.virt.hardware [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 791.960991] env[62753]: DEBUG nova.virt.hardware [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 791.961211] env[62753]: DEBUG nova.virt.hardware [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 791.961370] env[62753]: DEBUG nova.virt.hardware [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 791.961534] env[62753]: DEBUG nova.virt.hardware [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 791.961692] env[62753]: DEBUG nova.virt.hardware [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 791.961862] env[62753]: DEBUG nova.virt.hardware [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 791.962761] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c9078d7-e67c-4e02-940c-0d005adf0d05 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.973274] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c515bcd7-4e95-4dc0-976e-3b7417b2b901 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 791.987809] env[62753]: ERROR nova.compute.manager [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 257c7620-1533-42cc-b45b-b6615400c2d3, please check neutron logs for more information. [ 791.987809] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Traceback (most recent call last): [ 791.987809] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 791.987809] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] yield resources [ 791.987809] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 791.987809] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] self.driver.spawn(context, instance, image_meta, [ 791.987809] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 791.987809] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 791.987809] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 791.987809] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] vm_ref = self.build_virtual_machine(instance, [ 791.987809] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 791.988265] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] vif_infos = vmwarevif.get_vif_info(self._session, [ 791.988265] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 791.988265] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] for vif in network_info: [ 791.988265] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 791.988265] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] return self._sync_wrapper(fn, *args, **kwargs) [ 791.988265] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 791.988265] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] self.wait() [ 791.988265] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 791.988265] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] self[:] = self._gt.wait() [ 791.988265] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 791.988265] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] return self._exit_event.wait() [ 791.988265] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 791.988265] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] current.throw(*self._exc) [ 791.988683] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 791.988683] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] result = function(*args, **kwargs) [ 791.988683] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 791.988683] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] return func(*args, **kwargs) [ 791.988683] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 791.988683] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] raise e [ 791.988683] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 791.988683] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] nwinfo = self.network_api.allocate_for_instance( [ 791.988683] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 791.988683] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] created_port_ids = self._update_ports_for_instance( [ 791.988683] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 791.988683] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] with excutils.save_and_reraise_exception(): [ 791.988683] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 791.989108] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] self.force_reraise() [ 791.989108] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 791.989108] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] raise self.value [ 791.989108] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 791.989108] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] updated_port = self._update_port( [ 791.989108] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 791.989108] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] _ensure_no_port_binding_failure(port) [ 791.989108] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 791.989108] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] raise exception.PortBindingFailed(port_id=port['id']) [ 791.989108] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] nova.exception.PortBindingFailed: Binding failed for port 257c7620-1533-42cc-b45b-b6615400c2d3, please check neutron logs for more information. [ 791.989108] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] [ 791.989108] env[62753]: INFO nova.compute.manager [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Terminating instance [ 791.989764] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Acquiring lock "refresh_cache-e0c4a1d2-6a85-4826-88fa-d21556d744ac" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 792.164832] env[62753]: DEBUG nova.network.neutron [req-376e911e-a1af-46f2-a81f-77f672f39a83 req-7d50b731-f313-48bb-a4b2-f31bdf121f15 service nova] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.254980] env[62753]: DEBUG nova.network.neutron [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 792.325279] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 792.336065] env[62753]: DEBUG nova.network.neutron [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 792.510615] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-beafec06-a6bc-416e-8323-22d2673d4e33 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.518103] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abea8f18-baf5-4ac1-ad09-3d8295d466fa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.549045] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbb77988-763d-4ef6-91ea-ad60e111baed {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.556439] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73cd9ded-d30b-4191-a756-c16189cdbc13 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 792.569221] env[62753]: DEBUG nova.compute.provider_tree [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 792.667426] env[62753]: DEBUG oslo_concurrency.lockutils [req-376e911e-a1af-46f2-a81f-77f672f39a83 req-7d50b731-f313-48bb-a4b2-f31bdf121f15 service nova] Releasing lock "refresh_cache-e0c4a1d2-6a85-4826-88fa-d21556d744ac" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.667835] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Acquired lock "refresh_cache-e0c4a1d2-6a85-4826-88fa-d21556d744ac" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 792.668039] env[62753]: DEBUG nova.network.neutron [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 792.838962] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Releasing lock "refresh_cache-e260eaf7-98fe-461a-bc69-fc42dfcb79a8" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 792.839319] env[62753]: DEBUG nova.compute.manager [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 792.839466] env[62753]: DEBUG nova.compute.manager [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 792.839664] env[62753]: DEBUG nova.network.neutron [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 792.855364] env[62753]: DEBUG nova.network.neutron [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.072524] env[62753]: DEBUG nova.scheduler.client.report [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 793.190606] env[62753]: DEBUG nova.network.neutron [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 793.312197] env[62753]: DEBUG nova.network.neutron [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.357791] env[62753]: DEBUG nova.network.neutron [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 793.405358] env[62753]: DEBUG nova.compute.manager [req-2e8c8ac2-6d44-4c3b-ae9c-4ab1f0045610 req-a9563ced-bee3-4520-9bfe-128597f250d2 service nova] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Received event network-vif-deleted-257c7620-1533-42cc-b45b-b6615400c2d3 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 793.584983] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.860s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 793.585645] env[62753]: ERROR nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 43c09ea0-9e86-43c5-974f-b35a8f45f8b0, please check neutron logs for more information. [ 793.585645] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Traceback (most recent call last): [ 793.585645] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 793.585645] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] self.driver.spawn(context, instance, image_meta, [ 793.585645] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 793.585645] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] self._vmops.spawn(context, instance, image_meta, injected_files, [ 793.585645] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 793.585645] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] vm_ref = self.build_virtual_machine(instance, [ 793.585645] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 793.585645] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] vif_infos = vmwarevif.get_vif_info(self._session, [ 793.585645] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 793.585976] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] for vif in network_info: [ 793.585976] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 793.585976] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] return self._sync_wrapper(fn, *args, **kwargs) [ 793.585976] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 793.585976] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] self.wait() [ 793.585976] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 793.585976] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] self[:] = self._gt.wait() [ 793.585976] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 793.585976] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] return self._exit_event.wait() [ 793.585976] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 793.585976] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] result = hub.switch() [ 793.585976] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 793.585976] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] return self.greenlet.switch() [ 793.586337] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 793.586337] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] result = function(*args, **kwargs) [ 793.586337] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 793.586337] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] return func(*args, **kwargs) [ 793.586337] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 793.586337] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] raise e [ 793.586337] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 793.586337] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] nwinfo = self.network_api.allocate_for_instance( [ 793.586337] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 793.586337] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] created_port_ids = self._update_ports_for_instance( [ 793.586337] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 793.586337] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] with excutils.save_and_reraise_exception(): [ 793.586337] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 793.586865] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] self.force_reraise() [ 793.586865] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 793.586865] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] raise self.value [ 793.586865] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 793.586865] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] updated_port = self._update_port( [ 793.586865] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 793.586865] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] _ensure_no_port_binding_failure(port) [ 793.586865] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 793.586865] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] raise exception.PortBindingFailed(port_id=port['id']) [ 793.586865] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] nova.exception.PortBindingFailed: Binding failed for port 43c09ea0-9e86-43c5-974f-b35a8f45f8b0, please check neutron logs for more information. [ 793.586865] env[62753]: ERROR nova.compute.manager [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] [ 793.587196] env[62753]: DEBUG nova.compute.utils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Binding failed for port 43c09ea0-9e86-43c5-974f-b35a8f45f8b0, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 793.587596] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.757s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 793.587814] env[62753]: DEBUG nova.objects.instance [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Lazy-loading 'resources' on Instance uuid 8664ed5c-ecb9-4795-8499-31198cfd0450 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 793.589387] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Build of instance 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17 was re-scheduled: Binding failed for port 43c09ea0-9e86-43c5-974f-b35a8f45f8b0, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 793.589812] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 793.590069] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "refresh_cache-7831fd59-7a6d-4f6c-9f46-8cb1e2661c17" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 793.590271] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquired lock "refresh_cache-7831fd59-7a6d-4f6c-9f46-8cb1e2661c17" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 793.590406] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 793.814946] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Releasing lock "refresh_cache-e0c4a1d2-6a85-4826-88fa-d21556d744ac" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 793.815590] env[62753]: DEBUG nova.compute.manager [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 793.815714] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 793.815984] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-42ffa042-7cfa-4c3f-82a4-139b55cdfe9e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.826556] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b26f9fa-1c05-464f-833c-619b9dcca62f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 793.849288] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance e0c4a1d2-6a85-4826-88fa-d21556d744ac could not be found. [ 793.849599] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 793.851377] env[62753]: INFO nova.compute.manager [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Took 0.03 seconds to destroy the instance on the hypervisor. [ 793.851377] env[62753]: DEBUG oslo.service.loopingcall [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 793.851377] env[62753]: DEBUG nova.compute.manager [-] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 793.851377] env[62753]: DEBUG nova.network.neutron [-] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 793.860546] env[62753]: INFO nova.compute.manager [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] [instance: e260eaf7-98fe-461a-bc69-fc42dfcb79a8] Took 1.02 seconds to deallocate network for instance. [ 793.867734] env[62753]: DEBUG nova.network.neutron [-] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.112847] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.193274] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.358673] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-442bafb0-97a0-4287-95f7-9a068e2c7797 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.369092] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a42c8f9f-944c-4927-b9b1-15d928a98e76 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.372483] env[62753]: DEBUG nova.network.neutron [-] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 794.399482] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54bcfd19-5540-4f2d-9f81-f5ee52e66db3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.407154] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9868368-7ec7-42ef-9bc4-54d084bafe02 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 794.422314] env[62753]: DEBUG nova.compute.provider_tree [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 794.697109] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Releasing lock "refresh_cache-7831fd59-7a6d-4f6c-9f46-8cb1e2661c17" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 794.697109] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 794.697109] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 794.697109] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 794.713338] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 794.877067] env[62753]: INFO nova.compute.manager [-] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Took 1.03 seconds to deallocate network for instance. [ 794.878834] env[62753]: DEBUG nova.compute.claims [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 794.879011] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 794.896121] env[62753]: INFO nova.scheduler.client.report [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Deleted allocations for instance e260eaf7-98fe-461a-bc69-fc42dfcb79a8 [ 794.925533] env[62753]: DEBUG nova.scheduler.client.report [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 795.216134] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 795.406456] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8b4c55e-7264-4fc1-972b-1d4470e15a27 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388 tempest-FloatingIPsAssociationNegativeTestJSON-1012204388-project-member] Lock "e260eaf7-98fe-461a-bc69-fc42dfcb79a8" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 137.084s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.431810] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.844s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 795.435051] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.845s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 795.436618] env[62753]: INFO nova.compute.claims [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 795.456987] env[62753]: INFO nova.scheduler.client.report [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Deleted allocations for instance 8664ed5c-ecb9-4795-8499-31198cfd0450 [ 795.720662] env[62753]: INFO nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17] Took 1.02 seconds to deallocate network for instance. [ 795.909591] env[62753]: DEBUG nova.compute.manager [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 795.964055] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36781971-bcea-41c4-8274-711c9da7f02a tempest-ServersAaction247Test-311608892 tempest-ServersAaction247Test-311608892-project-member] Lock "8664ed5c-ecb9-4795-8499-31198cfd0450" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 23.861s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 796.436191] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 796.740984] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c67a2af5-7db5-4185-96e5-bc1e7022d7c3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.757249] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61baa32e-a238-4e3d-a4e2-5e8a544446a4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.807823] env[62753]: INFO nova.scheduler.client.report [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Deleted allocations for instance 7831fd59-7a6d-4f6c-9f46-8cb1e2661c17 [ 796.814267] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be3fc0d7-b770-4af2-a410-782236962437 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.823683] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efd308ad-afc2-48e5-b528-82594da4c27a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 796.840270] env[62753]: DEBUG nova.compute.provider_tree [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 797.322399] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "7831fd59-7a6d-4f6c-9f46-8cb1e2661c17" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 133.525s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.344301] env[62753]: DEBUG nova.scheduler.client.report [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 797.824832] env[62753]: DEBUG nova.compute.manager [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 797.851737] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.417s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 797.853041] env[62753]: DEBUG nova.compute.manager [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 797.855693] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.411s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 797.857189] env[62753]: INFO nova.compute.claims [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 798.348955] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.361921] env[62753]: DEBUG nova.compute.utils [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 798.364237] env[62753]: DEBUG nova.compute.manager [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Not allocating networking since 'none' was specified. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 798.685567] env[62753]: DEBUG oslo_concurrency.lockutils [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquiring lock "49311ef5-733a-4146-87ac-876a121a8d4d" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 798.685761] env[62753]: DEBUG oslo_concurrency.lockutils [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "49311ef5-733a-4146-87ac-876a121a8d4d" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 798.864836] env[62753]: DEBUG nova.compute.manager [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 799.113831] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd3bbd23-5964-46c9-bdfd-41d814524dc2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.125032] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46bc0d4e-0bc0-4a3f-83f8-4043c153ff88 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.155274] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29bd63ef-bfaa-48a6-8170-83e5b9db2766 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.164220] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf61a1cb-2eed-433a-8f12-2323cdc3842b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.177119] env[62753]: DEBUG nova.compute.provider_tree [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 799.680675] env[62753]: DEBUG nova.scheduler.client.report [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 799.880879] env[62753]: DEBUG nova.compute.manager [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 799.905837] env[62753]: DEBUG nova.virt.hardware [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 799.906099] env[62753]: DEBUG nova.virt.hardware [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 799.906256] env[62753]: DEBUG nova.virt.hardware [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 799.906473] env[62753]: DEBUG nova.virt.hardware [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 799.906634] env[62753]: DEBUG nova.virt.hardware [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 799.906783] env[62753]: DEBUG nova.virt.hardware [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 799.906992] env[62753]: DEBUG nova.virt.hardware [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 799.907169] env[62753]: DEBUG nova.virt.hardware [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 799.907333] env[62753]: DEBUG nova.virt.hardware [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 799.907493] env[62753]: DEBUG nova.virt.hardware [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 799.907662] env[62753]: DEBUG nova.virt.hardware [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 799.908605] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f24fc54-5852-4d1a-97cd-029467a41c8d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.916876] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47074a2f-2f46-42b5-aacd-8b860a01853b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.932290] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Instance VIF info [] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 799.937951] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Creating folder: Project (788265cd0b4345ffb2e25a360475b42c). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 799.938079] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53fddf3f-c452-4bd8-87a4-24e0faf43758 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.948978] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Created folder: Project (788265cd0b4345ffb2e25a360475b42c) in parent group-v284541. [ 799.949249] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Creating folder: Instances. Parent ref: group-v284552. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 799.949421] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-99102830-9837-4d79-9efd-ce571c504ecf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.960757] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Created folder: Instances in parent group-v284552. [ 799.960757] env[62753]: DEBUG oslo.service.loopingcall [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 799.960757] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 799.960896] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-7aacf780-70b3-49b4-b580-af3872f57aa7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 799.978019] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 799.978019] env[62753]: value = "task-1332124" [ 799.978019] env[62753]: _type = "Task" [ 799.978019] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 799.988525] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332124, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.185850] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.330s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 800.186249] env[62753]: DEBUG nova.compute.manager [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 800.188999] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.801s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 800.191195] env[62753]: INFO nova.compute.claims [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 800.488345] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332124, 'name': CreateVM_Task, 'duration_secs': 0.244514} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 800.488519] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 800.488934] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 800.489106] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 800.489430] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 800.490104] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e99791b4-e8b4-4d7b-b12f-ff2a996cd4c6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 800.495365] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 800.495365] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5293535c-757c-3d8c-05e4-7c831f69ee8a" [ 800.495365] env[62753]: _type = "Task" [ 800.495365] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 800.503409] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5293535c-757c-3d8c-05e4-7c831f69ee8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 800.695286] env[62753]: DEBUG nova.compute.utils [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 800.699356] env[62753]: DEBUG nova.compute.manager [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 800.699356] env[62753]: DEBUG nova.network.neutron [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 800.747810] env[62753]: DEBUG nova.policy [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6d8a39fb73cb49c7b839d18786f20603', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7957f17d217e46bcadec774cc0542d89', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 801.008333] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5293535c-757c-3d8c-05e4-7c831f69ee8a, 'name': SearchDatastore_Task, 'duration_secs': 0.007835} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.008633] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.008859] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 801.009099] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.009253] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.009415] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 801.009686] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0e0cf9e3-d40f-44ea-908b-898f6f71c334 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.018890] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 801.019088] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 801.019805] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-371fb092-22e3-425a-9a7d-14f7b01bc0a0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.025511] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 801.025511] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5201e291-4d5a-6fcc-49d6-92194e434a30" [ 801.025511] env[62753]: _type = "Task" [ 801.025511] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.034361] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5201e291-4d5a-6fcc-49d6-92194e434a30, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.039749] env[62753]: DEBUG nova.network.neutron [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Successfully created port: 04dda318-4cf1-44d4-bf27-fe6dce719865 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 801.202022] env[62753]: DEBUG nova.compute.manager [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 801.480519] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49b7aff6-f054-4500-b332-02ce44af6d47 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.488450] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-948ebdd0-5be6-4155-9955-b649d5147d8a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.532050] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d02e79b6-9182-47db-a15b-6e4f1ba5198f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.539883] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5201e291-4d5a-6fcc-49d6-92194e434a30, 'name': SearchDatastore_Task, 'duration_secs': 0.008715} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.542752] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6a435fcc-3788-4c09-b83a-0d413b3a8a10 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.546009] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06470797-2bd8-40af-8dfc-5ff3effc31d7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.559596] env[62753]: DEBUG nova.compute.provider_tree [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 801.562266] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 801.562266] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52b3c1ca-6d54-5b6a-701d-fb765a5f479c" [ 801.562266] env[62753]: _type = "Task" [ 801.562266] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.570357] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52b3c1ca-6d54-5b6a-701d-fb765a5f479c, 'name': SearchDatastore_Task, 'duration_secs': 0.009215} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 801.571250] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 801.571510] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 41476adb-95e5-4617-b178-9c981d13c43d/41476adb-95e5-4617-b178-9c981d13c43d.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 801.571762] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-58bf7acc-7271-4fa8-ba90-0ddf1065647f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 801.578189] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 801.578189] env[62753]: value = "task-1332129" [ 801.578189] env[62753]: _type = "Task" [ 801.578189] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 801.590791] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332129, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 801.743948] env[62753]: DEBUG nova.compute.manager [req-cbcb009d-903b-4015-b58b-3ba9644a64fb req-97e85ec4-f277-4280-9dde-5261d86403cf service nova] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Received event network-changed-04dda318-4cf1-44d4-bf27-fe6dce719865 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 801.744220] env[62753]: DEBUG nova.compute.manager [req-cbcb009d-903b-4015-b58b-3ba9644a64fb req-97e85ec4-f277-4280-9dde-5261d86403cf service nova] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Refreshing instance network info cache due to event network-changed-04dda318-4cf1-44d4-bf27-fe6dce719865. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 801.744418] env[62753]: DEBUG oslo_concurrency.lockutils [req-cbcb009d-903b-4015-b58b-3ba9644a64fb req-97e85ec4-f277-4280-9dde-5261d86403cf service nova] Acquiring lock "refresh_cache-3adb208d-cf70-4d93-b034-a3e87d029f90" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 801.744569] env[62753]: DEBUG oslo_concurrency.lockutils [req-cbcb009d-903b-4015-b58b-3ba9644a64fb req-97e85ec4-f277-4280-9dde-5261d86403cf service nova] Acquired lock "refresh_cache-3adb208d-cf70-4d93-b034-a3e87d029f90" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 801.745094] env[62753]: DEBUG nova.network.neutron [req-cbcb009d-903b-4015-b58b-3ba9644a64fb req-97e85ec4-f277-4280-9dde-5261d86403cf service nova] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Refreshing network info cache for port 04dda318-4cf1-44d4-bf27-fe6dce719865 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 801.971622] env[62753]: ERROR nova.compute.manager [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 04dda318-4cf1-44d4-bf27-fe6dce719865, please check neutron logs for more information. [ 801.971622] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 801.971622] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 801.971622] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 801.971622] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 801.971622] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 801.971622] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 801.971622] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 801.971622] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 801.971622] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 801.971622] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 801.971622] env[62753]: ERROR nova.compute.manager raise self.value [ 801.971622] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 801.971622] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 801.971622] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 801.971622] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 801.972193] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 801.972193] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 801.972193] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 04dda318-4cf1-44d4-bf27-fe6dce719865, please check neutron logs for more information. [ 801.972193] env[62753]: ERROR nova.compute.manager [ 801.972193] env[62753]: Traceback (most recent call last): [ 801.972193] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 801.972193] env[62753]: listener.cb(fileno) [ 801.972193] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 801.972193] env[62753]: result = function(*args, **kwargs) [ 801.972193] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 801.972193] env[62753]: return func(*args, **kwargs) [ 801.972193] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 801.972193] env[62753]: raise e [ 801.972193] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 801.972193] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 801.972193] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 801.972193] env[62753]: created_port_ids = self._update_ports_for_instance( [ 801.972193] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 801.972193] env[62753]: with excutils.save_and_reraise_exception(): [ 801.972193] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 801.972193] env[62753]: self.force_reraise() [ 801.972193] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 801.972193] env[62753]: raise self.value [ 801.972193] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 801.972193] env[62753]: updated_port = self._update_port( [ 801.972193] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 801.972193] env[62753]: _ensure_no_port_binding_failure(port) [ 801.972193] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 801.972193] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 801.973127] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 04dda318-4cf1-44d4-bf27-fe6dce719865, please check neutron logs for more information. [ 801.973127] env[62753]: Removing descriptor: 18 [ 802.064553] env[62753]: DEBUG nova.scheduler.client.report [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 802.092485] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332129, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.450701} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.092850] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 41476adb-95e5-4617-b178-9c981d13c43d/41476adb-95e5-4617-b178-9c981d13c43d.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 802.093113] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 802.093392] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-c1d03b8d-97de-4471-bc3b-404b1f8c1683 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.099769] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 802.099769] env[62753]: value = "task-1332130" [ 802.099769] env[62753]: _type = "Task" [ 802.099769] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.108564] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332130, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.213833] env[62753]: DEBUG nova.compute.manager [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 802.239157] env[62753]: DEBUG nova.virt.hardware [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 802.239472] env[62753]: DEBUG nova.virt.hardware [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 802.239639] env[62753]: DEBUG nova.virt.hardware [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 802.239823] env[62753]: DEBUG nova.virt.hardware [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 802.239967] env[62753]: DEBUG nova.virt.hardware [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 802.240130] env[62753]: DEBUG nova.virt.hardware [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 802.240344] env[62753]: DEBUG nova.virt.hardware [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 802.240496] env[62753]: DEBUG nova.virt.hardware [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 802.240664] env[62753]: DEBUG nova.virt.hardware [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 802.240823] env[62753]: DEBUG nova.virt.hardware [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 802.240990] env[62753]: DEBUG nova.virt.hardware [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 802.241870] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-725290a2-91ef-4f62-b50f-a6154de23a97 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.251435] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cef007d3-74cf-4d80-9750-49d8e0523d45 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.266421] env[62753]: ERROR nova.compute.manager [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 04dda318-4cf1-44d4-bf27-fe6dce719865, please check neutron logs for more information. [ 802.266421] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Traceback (most recent call last): [ 802.266421] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 802.266421] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] yield resources [ 802.266421] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 802.266421] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] self.driver.spawn(context, instance, image_meta, [ 802.266421] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 802.266421] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] self._vmops.spawn(context, instance, image_meta, injected_files, [ 802.266421] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 802.266421] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] vm_ref = self.build_virtual_machine(instance, [ 802.266421] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 802.266849] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] vif_infos = vmwarevif.get_vif_info(self._session, [ 802.266849] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 802.266849] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] for vif in network_info: [ 802.266849] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 802.266849] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] return self._sync_wrapper(fn, *args, **kwargs) [ 802.266849] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 802.266849] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] self.wait() [ 802.266849] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 802.266849] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] self[:] = self._gt.wait() [ 802.266849] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 802.266849] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] return self._exit_event.wait() [ 802.266849] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 802.266849] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] current.throw(*self._exc) [ 802.267309] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 802.267309] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] result = function(*args, **kwargs) [ 802.267309] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 802.267309] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] return func(*args, **kwargs) [ 802.267309] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 802.267309] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] raise e [ 802.267309] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 802.267309] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] nwinfo = self.network_api.allocate_for_instance( [ 802.267309] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 802.267309] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] created_port_ids = self._update_ports_for_instance( [ 802.267309] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 802.267309] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] with excutils.save_and_reraise_exception(): [ 802.267309] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 802.268088] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] self.force_reraise() [ 802.268088] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 802.268088] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] raise self.value [ 802.268088] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 802.268088] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] updated_port = self._update_port( [ 802.268088] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 802.268088] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] _ensure_no_port_binding_failure(port) [ 802.268088] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 802.268088] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] raise exception.PortBindingFailed(port_id=port['id']) [ 802.268088] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] nova.exception.PortBindingFailed: Binding failed for port 04dda318-4cf1-44d4-bf27-fe6dce719865, please check neutron logs for more information. [ 802.268088] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] [ 802.268088] env[62753]: INFO nova.compute.manager [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Terminating instance [ 802.268681] env[62753]: DEBUG nova.network.neutron [req-cbcb009d-903b-4015-b58b-3ba9644a64fb req-97e85ec4-f277-4280-9dde-5261d86403cf service nova] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 802.270412] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Acquiring lock "refresh_cache-3adb208d-cf70-4d93-b034-a3e87d029f90" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 802.342515] env[62753]: DEBUG nova.network.neutron [req-cbcb009d-903b-4015-b58b-3ba9644a64fb req-97e85ec4-f277-4280-9dde-5261d86403cf service nova] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 802.569702] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.381s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 802.570220] env[62753]: DEBUG nova.compute.manager [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 802.573124] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.286s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 802.574514] env[62753]: INFO nova.compute.claims [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 802.613112] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332130, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062964} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 802.613112] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 802.613112] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79bce910-35a5-40ff-a0b5-884cc5f16c00 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.632679] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Reconfiguring VM instance instance-00000030 to attach disk [datastore2] 41476adb-95e5-4617-b178-9c981d13c43d/41476adb-95e5-4617-b178-9c981d13c43d.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 802.633189] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8389b360-8550-4db2-8b6d-d421fc707320 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 802.652510] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 802.652510] env[62753]: value = "task-1332131" [ 802.652510] env[62753]: _type = "Task" [ 802.652510] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 802.660087] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332131, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 802.846345] env[62753]: DEBUG oslo_concurrency.lockutils [req-cbcb009d-903b-4015-b58b-3ba9644a64fb req-97e85ec4-f277-4280-9dde-5261d86403cf service nova] Releasing lock "refresh_cache-3adb208d-cf70-4d93-b034-a3e87d029f90" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 802.846767] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Acquired lock "refresh_cache-3adb208d-cf70-4d93-b034-a3e87d029f90" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 802.846952] env[62753]: DEBUG nova.network.neutron [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 803.079030] env[62753]: DEBUG nova.compute.utils [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 803.082367] env[62753]: DEBUG nova.compute.manager [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 803.083139] env[62753]: DEBUG nova.network.neutron [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 803.138115] env[62753]: DEBUG nova.policy [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd86dfa341c7a41938421b65e5d8d35b8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'fdd163b294804fffbf0e92873611d5ea', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 803.162288] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332131, 'name': ReconfigVM_Task, 'duration_secs': 0.240596} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.162567] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Reconfigured VM instance instance-00000030 to attach disk [datastore2] 41476adb-95e5-4617-b178-9c981d13c43d/41476adb-95e5-4617-b178-9c981d13c43d.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 803.163171] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-495c883e-f7cc-424e-87ed-0c8a158e0750 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.171171] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 803.171171] env[62753]: value = "task-1332132" [ 803.171171] env[62753]: _type = "Task" [ 803.171171] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.177185] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332132, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.370842] env[62753]: DEBUG nova.network.neutron [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 803.384234] env[62753]: DEBUG nova.network.neutron [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Successfully created port: f362843b-39bc-4173-8881-61ed1caca7c6 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 803.487020] env[62753]: DEBUG nova.network.neutron [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 803.583786] env[62753]: DEBUG nova.compute.manager [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 803.679125] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332132, 'name': Rename_Task, 'duration_secs': 0.122903} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 803.679466] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 803.683555] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c065f128-c307-40d3-b2d0-d799054fe389 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.690679] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 803.690679] env[62753]: value = "task-1332133" [ 803.690679] env[62753]: _type = "Task" [ 803.690679] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 803.709869] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332133, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 803.779669] env[62753]: DEBUG nova.compute.manager [req-fe54b6b4-3935-4889-8826-cf9403f1d0ce req-2620d38b-f28b-4570-863b-497a6cdf4773 service nova] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Received event network-vif-deleted-04dda318-4cf1-44d4-bf27-fe6dce719865 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 803.873193] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44d80484-3f33-4760-8cae-a4d598f9a05b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.880434] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ece488a1-5f48-4ba9-aa50-68230df6840c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.914726] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e782ca8-413b-45af-a629-e01cd2527a36 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.922055] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ae29de9-9109-44da-a023-17b87ba91c26 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 803.938589] env[62753]: DEBUG nova.compute.provider_tree [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 803.989926] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Releasing lock "refresh_cache-3adb208d-cf70-4d93-b034-a3e87d029f90" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 803.990476] env[62753]: DEBUG nova.compute.manager [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 803.998014] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 803.998014] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5abee882-59f0-48d0-a95a-9e3f180a125e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.006162] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4d0a142-e04f-43c3-a1b6-cad1131c64fa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.029880] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 3adb208d-cf70-4d93-b034-a3e87d029f90 could not be found. [ 804.031018] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 804.031018] env[62753]: INFO nova.compute.manager [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Took 0.04 seconds to destroy the instance on the hypervisor. [ 804.031018] env[62753]: DEBUG oslo.service.loopingcall [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 804.031018] env[62753]: DEBUG nova.compute.manager [-] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 804.031018] env[62753]: DEBUG nova.network.neutron [-] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 804.046820] env[62753]: DEBUG nova.network.neutron [-] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 804.202115] env[62753]: DEBUG oslo_vmware.api [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332133, 'name': PowerOnVM_Task, 'duration_secs': 0.447771} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 804.202404] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 804.202596] env[62753]: INFO nova.compute.manager [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Took 4.32 seconds to spawn the instance on the hypervisor. [ 804.202761] env[62753]: DEBUG nova.compute.manager [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 804.203728] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c39dd8a2-b7b1-44de-ab10-940e1a3b27c6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.294657] env[62753]: ERROR nova.compute.manager [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port f362843b-39bc-4173-8881-61ed1caca7c6, please check neutron logs for more information. [ 804.294657] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 804.294657] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 804.294657] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 804.294657] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.294657] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 804.294657] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.294657] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 804.294657] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.294657] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 804.294657] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.294657] env[62753]: ERROR nova.compute.manager raise self.value [ 804.294657] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.294657] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 804.294657] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.294657] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 804.295432] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.295432] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 804.295432] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port f362843b-39bc-4173-8881-61ed1caca7c6, please check neutron logs for more information. [ 804.295432] env[62753]: ERROR nova.compute.manager [ 804.295432] env[62753]: Traceback (most recent call last): [ 804.295432] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 804.295432] env[62753]: listener.cb(fileno) [ 804.295432] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 804.295432] env[62753]: result = function(*args, **kwargs) [ 804.295432] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 804.295432] env[62753]: return func(*args, **kwargs) [ 804.295432] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 804.295432] env[62753]: raise e [ 804.295432] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 804.295432] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 804.295432] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.295432] env[62753]: created_port_ids = self._update_ports_for_instance( [ 804.295432] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.295432] env[62753]: with excutils.save_and_reraise_exception(): [ 804.295432] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.295432] env[62753]: self.force_reraise() [ 804.295432] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.295432] env[62753]: raise self.value [ 804.295432] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.295432] env[62753]: updated_port = self._update_port( [ 804.295432] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.295432] env[62753]: _ensure_no_port_binding_failure(port) [ 804.295432] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.295432] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 804.296334] env[62753]: nova.exception.PortBindingFailed: Binding failed for port f362843b-39bc-4173-8881-61ed1caca7c6, please check neutron logs for more information. [ 804.296334] env[62753]: Removing descriptor: 18 [ 804.442199] env[62753]: DEBUG nova.scheduler.client.report [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 804.549832] env[62753]: DEBUG nova.network.neutron [-] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 804.596535] env[62753]: DEBUG nova.compute.manager [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 804.622169] env[62753]: DEBUG nova.virt.hardware [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:28:33Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='3099e560-a981-4712-afe9-8121de1ddcac',id=38,is_public=True,memory_mb=192,name='tempest-test_resize_flavor_-1416542554',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 804.622490] env[62753]: DEBUG nova.virt.hardware [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 804.622602] env[62753]: DEBUG nova.virt.hardware [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 804.623356] env[62753]: DEBUG nova.virt.hardware [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 804.623356] env[62753]: DEBUG nova.virt.hardware [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 804.623356] env[62753]: DEBUG nova.virt.hardware [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 804.623356] env[62753]: DEBUG nova.virt.hardware [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 804.623605] env[62753]: DEBUG nova.virt.hardware [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 804.623605] env[62753]: DEBUG nova.virt.hardware [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 804.623748] env[62753]: DEBUG nova.virt.hardware [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 804.623924] env[62753]: DEBUG nova.virt.hardware [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 804.624877] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d9f6e97-6c19-4798-8a2f-eaeae0729119 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.632698] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5b31eab-9130-40df-8a56-aa5f06b33cff {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 804.646377] env[62753]: ERROR nova.compute.manager [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port f362843b-39bc-4173-8881-61ed1caca7c6, please check neutron logs for more information. [ 804.646377] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Traceback (most recent call last): [ 804.646377] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 804.646377] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] yield resources [ 804.646377] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 804.646377] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] self.driver.spawn(context, instance, image_meta, [ 804.646377] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 804.646377] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] self._vmops.spawn(context, instance, image_meta, injected_files, [ 804.646377] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 804.646377] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] vm_ref = self.build_virtual_machine(instance, [ 804.646377] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 804.646906] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] vif_infos = vmwarevif.get_vif_info(self._session, [ 804.646906] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 804.646906] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] for vif in network_info: [ 804.646906] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 804.646906] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] return self._sync_wrapper(fn, *args, **kwargs) [ 804.646906] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 804.646906] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] self.wait() [ 804.646906] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 804.646906] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] self[:] = self._gt.wait() [ 804.646906] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 804.646906] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] return self._exit_event.wait() [ 804.646906] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 804.646906] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] current.throw(*self._exc) [ 804.647621] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 804.647621] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] result = function(*args, **kwargs) [ 804.647621] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 804.647621] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] return func(*args, **kwargs) [ 804.647621] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 804.647621] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] raise e [ 804.647621] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 804.647621] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] nwinfo = self.network_api.allocate_for_instance( [ 804.647621] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 804.647621] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] created_port_ids = self._update_ports_for_instance( [ 804.647621] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 804.647621] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] with excutils.save_and_reraise_exception(): [ 804.647621] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 804.648616] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] self.force_reraise() [ 804.648616] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 804.648616] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] raise self.value [ 804.648616] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 804.648616] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] updated_port = self._update_port( [ 804.648616] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 804.648616] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] _ensure_no_port_binding_failure(port) [ 804.648616] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 804.648616] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] raise exception.PortBindingFailed(port_id=port['id']) [ 804.648616] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] nova.exception.PortBindingFailed: Binding failed for port f362843b-39bc-4173-8881-61ed1caca7c6, please check neutron logs for more information. [ 804.648616] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] [ 804.648616] env[62753]: INFO nova.compute.manager [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Terminating instance [ 804.649028] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Acquiring lock "refresh_cache-f63a8912-0abf-4c76-aaae-4085b3909688" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 804.649028] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Acquired lock "refresh_cache-f63a8912-0abf-4c76-aaae-4085b3909688" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 804.649108] env[62753]: DEBUG nova.network.neutron [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 804.723031] env[62753]: INFO nova.compute.manager [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Took 27.17 seconds to build instance. [ 804.950063] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 804.950063] env[62753]: DEBUG nova.compute.manager [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 804.953025] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 19.126s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 805.053140] env[62753]: INFO nova.compute.manager [-] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Took 1.02 seconds to deallocate network for instance. [ 805.055801] env[62753]: DEBUG nova.compute.claims [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 805.056051] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 805.167869] env[62753]: DEBUG nova.network.neutron [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.224922] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea11b5b5-7a84-45cb-923b-c5d2f3e9b6f1 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Lock "41476adb-95e5-4617-b178-9c981d13c43d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 121.947s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 805.225928] env[62753]: DEBUG nova.network.neutron [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 805.453551] env[62753]: DEBUG nova.compute.utils [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 805.455437] env[62753]: DEBUG nova.compute.manager [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 805.455960] env[62753]: DEBUG nova.network.neutron [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 805.462326] env[62753]: INFO nova.compute.manager [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Rebuilding instance [ 805.496489] env[62753]: DEBUG nova.policy [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae52823f7c93454e8089261c3ca44321', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d52d6eaee934be5ab0e0003df1ce316', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 805.518368] env[62753]: DEBUG nova.compute.manager [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 805.519300] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1483fe7-6b05-441e-bf4d-b6f97ecdd9c4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.728968] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Releasing lock "refresh_cache-f63a8912-0abf-4c76-aaae-4085b3909688" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 805.729628] env[62753]: DEBUG nova.compute.manager [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 805.729848] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 805.730263] env[62753]: DEBUG nova.compute.manager [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 805.734524] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ceea2b36-b123-4a30-ada1-de982abdb4a7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.750552] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-475e8f1b-26a3-4fcf-9cda-db75982ecc93 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 805.779061] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance f63a8912-0abf-4c76-aaae-4085b3909688 could not be found. [ 805.779336] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 805.779558] env[62753]: INFO nova.compute.manager [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Took 0.05 seconds to destroy the instance on the hypervisor. [ 805.780123] env[62753]: DEBUG oslo.service.loopingcall [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 805.780369] env[62753]: DEBUG nova.compute.manager [-] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 805.780507] env[62753]: DEBUG nova.network.neutron [-] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 805.806406] env[62753]: DEBUG nova.network.neutron [-] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 805.809672] env[62753]: DEBUG nova.compute.manager [req-b7ad1893-ab9c-4925-b41a-82140f3aa4a1 req-759e22cf-01a8-44b7-aaf7-647e25ec26df service nova] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Received event network-changed-f362843b-39bc-4173-8881-61ed1caca7c6 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 805.809853] env[62753]: DEBUG nova.compute.manager [req-b7ad1893-ab9c-4925-b41a-82140f3aa4a1 req-759e22cf-01a8-44b7-aaf7-647e25ec26df service nova] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Refreshing instance network info cache due to event network-changed-f362843b-39bc-4173-8881-61ed1caca7c6. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 805.810231] env[62753]: DEBUG oslo_concurrency.lockutils [req-b7ad1893-ab9c-4925-b41a-82140f3aa4a1 req-759e22cf-01a8-44b7-aaf7-647e25ec26df service nova] Acquiring lock "refresh_cache-f63a8912-0abf-4c76-aaae-4085b3909688" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 805.810231] env[62753]: DEBUG oslo_concurrency.lockutils [req-b7ad1893-ab9c-4925-b41a-82140f3aa4a1 req-759e22cf-01a8-44b7-aaf7-647e25ec26df service nova] Acquired lock "refresh_cache-f63a8912-0abf-4c76-aaae-4085b3909688" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 805.810386] env[62753]: DEBUG nova.network.neutron [req-b7ad1893-ab9c-4925-b41a-82140f3aa4a1 req-759e22cf-01a8-44b7-aaf7-647e25ec26df service nova] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Refreshing network info cache for port f362843b-39bc-4173-8881-61ed1caca7c6 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 805.833151] env[62753]: DEBUG nova.network.neutron [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Successfully created port: 33143ac6-b800-4c58-b336-0db7572c50a7 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 805.958996] env[62753]: DEBUG nova.compute.manager [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 805.989443] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 805.989621] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance e0c4a1d2-6a85-4826-88fa-d21556d744ac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 805.989751] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 41476adb-95e5-4617-b178-9c981d13c43d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 805.989970] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 3adb208d-cf70-4d93-b034-a3e87d029f90 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 805.990076] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance f63a8912-0abf-4c76-aaae-4085b3909688 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 805.990181] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 7563bf4e-39c5-4831-a415-cb311e237762 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 806.031424] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 806.031788] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-51344ff5-e741-4279-ab14-f051f25733e8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.040016] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 806.040016] env[62753]: value = "task-1332135" [ 806.040016] env[62753]: _type = "Task" [ 806.040016] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.048635] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332135, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.256585] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.312547] env[62753]: DEBUG nova.network.neutron [-] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.333246] env[62753]: DEBUG nova.network.neutron [req-b7ad1893-ab9c-4925-b41a-82140f3aa4a1 req-759e22cf-01a8-44b7-aaf7-647e25ec26df service nova] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 806.407695] env[62753]: DEBUG nova.network.neutron [req-b7ad1893-ab9c-4925-b41a-82140f3aa4a1 req-759e22cf-01a8-44b7-aaf7-647e25ec26df service nova] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 806.493414] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 806.549794] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332135, 'name': PowerOffVM_Task, 'duration_secs': 0.184566} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 806.550163] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 806.550424] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 806.551215] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10665e4a-77af-4e6f-9b7d-de31ba200af7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.559259] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 806.559495] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-93e0f7b1-737d-46dd-bbd5-0d991a11cae0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.586951] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 806.587179] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 806.587526] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Deleting the datastore file [datastore2] 41476adb-95e5-4617-b178-9c981d13c43d {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 806.587672] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d9302e83-43f8-4cdf-8af6-63246a22257d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 806.593736] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 806.593736] env[62753]: value = "task-1332138" [ 806.593736] env[62753]: _type = "Task" [ 806.593736] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 806.601954] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332138, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 806.672927] env[62753]: ERROR nova.compute.manager [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 33143ac6-b800-4c58-b336-0db7572c50a7, please check neutron logs for more information. [ 806.672927] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 806.672927] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 806.672927] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 806.672927] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 806.672927] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 806.672927] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 806.672927] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 806.672927] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 806.672927] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 806.672927] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 806.672927] env[62753]: ERROR nova.compute.manager raise self.value [ 806.672927] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 806.672927] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 806.672927] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 806.672927] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 806.673890] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 806.673890] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 806.673890] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 33143ac6-b800-4c58-b336-0db7572c50a7, please check neutron logs for more information. [ 806.673890] env[62753]: ERROR nova.compute.manager [ 806.673890] env[62753]: Traceback (most recent call last): [ 806.673890] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 806.673890] env[62753]: listener.cb(fileno) [ 806.673890] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 806.673890] env[62753]: result = function(*args, **kwargs) [ 806.673890] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 806.673890] env[62753]: return func(*args, **kwargs) [ 806.673890] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 806.673890] env[62753]: raise e [ 806.673890] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 806.673890] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 806.673890] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 806.673890] env[62753]: created_port_ids = self._update_ports_for_instance( [ 806.673890] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 806.673890] env[62753]: with excutils.save_and_reraise_exception(): [ 806.673890] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 806.673890] env[62753]: self.force_reraise() [ 806.673890] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 806.673890] env[62753]: raise self.value [ 806.673890] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 806.673890] env[62753]: updated_port = self._update_port( [ 806.673890] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 806.673890] env[62753]: _ensure_no_port_binding_failure(port) [ 806.673890] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 806.673890] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 806.675123] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 33143ac6-b800-4c58-b336-0db7572c50a7, please check neutron logs for more information. [ 806.675123] env[62753]: Removing descriptor: 18 [ 806.816643] env[62753]: INFO nova.compute.manager [-] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Took 1.04 seconds to deallocate network for instance. [ 806.818973] env[62753]: DEBUG nova.compute.claims [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 806.819173] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 806.910971] env[62753]: DEBUG oslo_concurrency.lockutils [req-b7ad1893-ab9c-4925-b41a-82140f3aa4a1 req-759e22cf-01a8-44b7-aaf7-647e25ec26df service nova] Releasing lock "refresh_cache-f63a8912-0abf-4c76-aaae-4085b3909688" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 806.910971] env[62753]: DEBUG nova.compute.manager [req-b7ad1893-ab9c-4925-b41a-82140f3aa4a1 req-759e22cf-01a8-44b7-aaf7-647e25ec26df service nova] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Received event network-vif-deleted-f362843b-39bc-4173-8881-61ed1caca7c6 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 806.968379] env[62753]: DEBUG nova.compute.manager [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 806.994622] env[62753]: DEBUG nova.virt.hardware [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 806.994887] env[62753]: DEBUG nova.virt.hardware [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 806.995067] env[62753]: DEBUG nova.virt.hardware [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 806.995274] env[62753]: DEBUG nova.virt.hardware [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 806.995422] env[62753]: DEBUG nova.virt.hardware [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 806.995567] env[62753]: DEBUG nova.virt.hardware [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 806.995778] env[62753]: DEBUG nova.virt.hardware [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 806.995933] env[62753]: DEBUG nova.virt.hardware [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 806.996122] env[62753]: DEBUG nova.virt.hardware [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 806.996287] env[62753]: DEBUG nova.virt.hardware [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 806.996456] env[62753]: DEBUG nova.virt.hardware [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 806.997213] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 67f5d96c-b870-442a-a0f3-a70533c045ea has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 806.998990] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b3586f5-0481-44d5-b6c7-a25403043e16 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.007677] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd433a7c-294b-4ac2-826b-0b77cb5c68f5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 807.022896] env[62753]: ERROR nova.compute.manager [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 33143ac6-b800-4c58-b336-0db7572c50a7, please check neutron logs for more information. [ 807.022896] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Traceback (most recent call last): [ 807.022896] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 807.022896] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] yield resources [ 807.022896] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 807.022896] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] self.driver.spawn(context, instance, image_meta, [ 807.022896] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 807.022896] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] self._vmops.spawn(context, instance, image_meta, injected_files, [ 807.022896] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 807.022896] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] vm_ref = self.build_virtual_machine(instance, [ 807.022896] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 807.023418] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] vif_infos = vmwarevif.get_vif_info(self._session, [ 807.023418] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 807.023418] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] for vif in network_info: [ 807.023418] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 807.023418] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] return self._sync_wrapper(fn, *args, **kwargs) [ 807.023418] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 807.023418] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] self.wait() [ 807.023418] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 807.023418] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] self[:] = self._gt.wait() [ 807.023418] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 807.023418] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] return self._exit_event.wait() [ 807.023418] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 807.023418] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] current.throw(*self._exc) [ 807.023854] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 807.023854] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] result = function(*args, **kwargs) [ 807.023854] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 807.023854] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] return func(*args, **kwargs) [ 807.023854] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 807.023854] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] raise e [ 807.023854] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 807.023854] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] nwinfo = self.network_api.allocate_for_instance( [ 807.023854] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 807.023854] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] created_port_ids = self._update_ports_for_instance( [ 807.023854] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 807.023854] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] with excutils.save_and_reraise_exception(): [ 807.023854] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 807.024311] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] self.force_reraise() [ 807.024311] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 807.024311] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] raise self.value [ 807.024311] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 807.024311] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] updated_port = self._update_port( [ 807.024311] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 807.024311] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] _ensure_no_port_binding_failure(port) [ 807.024311] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 807.024311] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] raise exception.PortBindingFailed(port_id=port['id']) [ 807.024311] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] nova.exception.PortBindingFailed: Binding failed for port 33143ac6-b800-4c58-b336-0db7572c50a7, please check neutron logs for more information. [ 807.024311] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] [ 807.024311] env[62753]: INFO nova.compute.manager [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Terminating instance [ 807.025111] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "refresh_cache-7563bf4e-39c5-4831-a415-cb311e237762" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 807.025274] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "refresh_cache-7563bf4e-39c5-4831-a415-cb311e237762" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 807.025437] env[62753]: DEBUG nova.network.neutron [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 807.103715] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332138, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.190064} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 807.103946] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 807.104142] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 807.104354] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 807.504556] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 92e86d6a-e455-4984-9fdb-d1cdf856e729 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 807.544613] env[62753]: DEBUG nova.network.neutron [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 807.618253] env[62753]: DEBUG nova.network.neutron [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 807.836025] env[62753]: DEBUG nova.compute.manager [req-394eb3cc-8e28-449b-9ccf-e0575fcbb995 req-8a8b9f27-4f0d-442e-bb76-00e656fe65d3 service nova] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Received event network-changed-33143ac6-b800-4c58-b336-0db7572c50a7 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 807.836229] env[62753]: DEBUG nova.compute.manager [req-394eb3cc-8e28-449b-9ccf-e0575fcbb995 req-8a8b9f27-4f0d-442e-bb76-00e656fe65d3 service nova] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Refreshing instance network info cache due to event network-changed-33143ac6-b800-4c58-b336-0db7572c50a7. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 807.836423] env[62753]: DEBUG oslo_concurrency.lockutils [req-394eb3cc-8e28-449b-9ccf-e0575fcbb995 req-8a8b9f27-4f0d-442e-bb76-00e656fe65d3 service nova] Acquiring lock "refresh_cache-7563bf4e-39c5-4831-a415-cb311e237762" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.008057] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 7931ae4b-c486-4e57-9bf5-14166a1d4201 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 808.121061] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "refresh_cache-7563bf4e-39c5-4831-a415-cb311e237762" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 808.121457] env[62753]: DEBUG nova.compute.manager [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 808.121645] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 808.121958] env[62753]: DEBUG oslo_concurrency.lockutils [req-394eb3cc-8e28-449b-9ccf-e0575fcbb995 req-8a8b9f27-4f0d-442e-bb76-00e656fe65d3 service nova] Acquired lock "refresh_cache-7563bf4e-39c5-4831-a415-cb311e237762" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.122140] env[62753]: DEBUG nova.network.neutron [req-394eb3cc-8e28-449b-9ccf-e0575fcbb995 req-8a8b9f27-4f0d-442e-bb76-00e656fe65d3 service nova] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Refreshing network info cache for port 33143ac6-b800-4c58-b336-0db7572c50a7 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 808.123475] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4cb02514-a628-4966-98dd-dd517d728b79 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.133853] env[62753]: DEBUG nova.virt.hardware [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 808.134080] env[62753]: DEBUG nova.virt.hardware [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 808.134239] env[62753]: DEBUG nova.virt.hardware [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 808.134414] env[62753]: DEBUG nova.virt.hardware [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 808.134559] env[62753]: DEBUG nova.virt.hardware [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 808.134703] env[62753]: DEBUG nova.virt.hardware [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 808.134905] env[62753]: DEBUG nova.virt.hardware [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 808.135074] env[62753]: DEBUG nova.virt.hardware [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 808.135245] env[62753]: DEBUG nova.virt.hardware [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 808.135405] env[62753]: DEBUG nova.virt.hardware [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 808.135592] env[62753]: DEBUG nova.virt.hardware [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 808.138464] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d4dedf2-2309-4382-b7ab-91acc0850cad {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.149190] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd45d359-4d95-43c7-99f7-1a5747d679f4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.156872] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ee18d43-eeb9-4541-906f-578c4943ead0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.165166] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7563bf4e-39c5-4831-a415-cb311e237762 could not be found. [ 808.165359] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 808.165529] env[62753]: INFO nova.compute.manager [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Took 0.04 seconds to destroy the instance on the hypervisor. [ 808.165749] env[62753]: DEBUG oslo.service.loopingcall [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.166262] env[62753]: DEBUG nova.compute.manager [-] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 808.166358] env[62753]: DEBUG nova.network.neutron [-] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 808.175365] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Instance VIF info [] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 808.180721] env[62753]: DEBUG oslo.service.loopingcall [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 808.181347] env[62753]: DEBUG nova.network.neutron [-] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.182396] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 808.182937] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e72bc73a-2ee6-4423-a62c-634909e67ef0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.199431] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 808.199431] env[62753]: value = "task-1332139" [ 808.199431] env[62753]: _type = "Task" [ 808.199431] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.206363] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332139, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 808.511187] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 16c82d55-1420-4f83-a547-295e793de9df has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 808.641416] env[62753]: DEBUG nova.network.neutron [req-394eb3cc-8e28-449b-9ccf-e0575fcbb995 req-8a8b9f27-4f0d-442e-bb76-00e656fe65d3 service nova] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 808.684715] env[62753]: DEBUG nova.network.neutron [-] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.710638] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332139, 'name': CreateVM_Task, 'duration_secs': 0.243438} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 808.710820] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 808.711273] env[62753]: DEBUG oslo_concurrency.lockutils [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 808.711391] env[62753]: DEBUG oslo_concurrency.lockutils [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 808.711746] env[62753]: DEBUG oslo_concurrency.lockutils [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 808.711943] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f776430d-c45d-4210-b074-c151ec175ac9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 808.714278] env[62753]: DEBUG nova.network.neutron [req-394eb3cc-8e28-449b-9ccf-e0575fcbb995 req-8a8b9f27-4f0d-442e-bb76-00e656fe65d3 service nova] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 808.719021] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 808.719021] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f82d76-709f-0fbb-f128-a3e380cb7c01" [ 808.719021] env[62753]: _type = "Task" [ 808.719021] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 808.726598] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f82d76-709f-0fbb-f128-a3e380cb7c01, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.015695] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 3bc05fdb-81be-4764-b9d9-01acd125b020 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 809.186666] env[62753]: INFO nova.compute.manager [-] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Took 1.02 seconds to deallocate network for instance. [ 809.188975] env[62753]: DEBUG nova.compute.claims [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 809.189200] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 809.216865] env[62753]: DEBUG oslo_concurrency.lockutils [req-394eb3cc-8e28-449b-9ccf-e0575fcbb995 req-8a8b9f27-4f0d-442e-bb76-00e656fe65d3 service nova] Releasing lock "refresh_cache-7563bf4e-39c5-4831-a415-cb311e237762" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.217105] env[62753]: DEBUG nova.compute.manager [req-394eb3cc-8e28-449b-9ccf-e0575fcbb995 req-8a8b9f27-4f0d-442e-bb76-00e656fe65d3 service nova] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Received event network-vif-deleted-33143ac6-b800-4c58-b336-0db7572c50a7 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 809.228746] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f82d76-709f-0fbb-f128-a3e380cb7c01, 'name': SearchDatastore_Task, 'duration_secs': 0.008543} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.229010] env[62753]: DEBUG oslo_concurrency.lockutils [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 809.229232] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 809.229451] env[62753]: DEBUG oslo_concurrency.lockutils [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 809.229602] env[62753]: DEBUG oslo_concurrency.lockutils [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 809.229787] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 809.230018] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7427cfc4-1ef1-42d3-80fc-91fc24e0b280 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.237402] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 809.237566] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 809.238231] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ab46a025-bd2d-436d-87aa-b0a6838d189e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.242682] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 809.242682] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52efa848-da3e-6957-57a2-5e69919cafb8" [ 809.242682] env[62753]: _type = "Task" [ 809.242682] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.249654] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52efa848-da3e-6957-57a2-5e69919cafb8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 809.518391] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 046fc3dc-a187-4fda-bc66-345e1226f83d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 809.753796] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52efa848-da3e-6957-57a2-5e69919cafb8, 'name': SearchDatastore_Task, 'duration_secs': 0.026559} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 809.754606] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a23b049e-f6c2-4900-8a4d-49272ccf4892 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 809.759716] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 809.759716] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c44cbe-88df-03d4-ee31-03e2647347ac" [ 809.759716] env[62753]: _type = "Task" [ 809.759716] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 809.767873] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c44cbe-88df-03d4-ee31-03e2647347ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.021626] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 810.270147] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c44cbe-88df-03d4-ee31-03e2647347ac, 'name': SearchDatastore_Task, 'duration_secs': 0.008922} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.270405] env[62753]: DEBUG oslo_concurrency.lockutils [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 810.271076] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 41476adb-95e5-4617-b178-9c981d13c43d/41476adb-95e5-4617-b178-9c981d13c43d.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 810.271076] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-220ff5ee-7aa5-45e8-90aa-03b9fb513538 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.277432] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 810.277432] env[62753]: value = "task-1332141" [ 810.277432] env[62753]: _type = "Task" [ 810.277432] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.284736] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332141, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 810.525641] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance f132230d-0fba-4293-b940-85cc7ef1cad1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 810.787505] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332141, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.493356} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 810.787505] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 41476adb-95e5-4617-b178-9c981d13c43d/41476adb-95e5-4617-b178-9c981d13c43d.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 810.787695] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 810.787908] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-82d3c29f-1e70-4f50-af89-24d33963b89b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 810.794118] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 810.794118] env[62753]: value = "task-1332143" [ 810.794118] env[62753]: _type = "Task" [ 810.794118] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 810.801526] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332143, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.029309] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance a6ee010c-6305-4009-80e0-92a2c58bcd7b has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 811.304352] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332143, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086342} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.304559] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 811.305346] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdc85964-009e-4943-a250-5eb807c95f14 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.324985] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] 41476adb-95e5-4617-b178-9c981d13c43d/41476adb-95e5-4617-b178-9c981d13c43d.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 811.325232] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-25f56bf9-dbc2-4a22-a77d-6b93f4e39ef9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.343146] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 811.343146] env[62753]: value = "task-1332144" [ 811.343146] env[62753]: _type = "Task" [ 811.343146] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.350326] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332144, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 811.532718] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 811.852835] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332144, 'name': ReconfigVM_Task, 'duration_secs': 0.300571} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 811.853113] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Reconfigured VM instance instance-00000030 to attach disk [datastore1] 41476adb-95e5-4617-b178-9c981d13c43d/41476adb-95e5-4617-b178-9c981d13c43d.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 811.853731] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0d81b375-729f-4a29-aed0-4f23a8cbb545 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 811.859662] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 811.859662] env[62753]: value = "task-1332145" [ 811.859662] env[62753]: _type = "Task" [ 811.859662] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 811.869532] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332145, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.036202] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 49311ef5-733a-4146-87ac-876a121a8d4d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 812.036498] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 812.036601] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 812.236602] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d8c1bcf-46f0-43b9-96e0-d25eb37713c2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.243958] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a38b5e66-db45-47fd-9b0c-fb09311cef5a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.273335] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f31904a-0935-49ba-bc79-589fc7d0c913 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.280161] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b6b8186-148f-43ee-bd2e-6de523fcffd7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.292766] env[62753]: DEBUG nova.compute.provider_tree [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 812.368971] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332145, 'name': Rename_Task, 'duration_secs': 0.128468} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.369216] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 812.369482] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-047559b5-7c24-44da-99cf-94b382b63e54 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 812.375204] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 812.375204] env[62753]: value = "task-1332146" [ 812.375204] env[62753]: _type = "Task" [ 812.375204] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 812.385626] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332146, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 812.796908] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 812.885430] env[62753]: DEBUG oslo_vmware.api [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332146, 'name': PowerOnVM_Task, 'duration_secs': 0.443821} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 812.885709] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 812.885908] env[62753]: DEBUG nova.compute.manager [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 812.886663] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8af35ca5-8ac0-45da-9084-6bddc33fd68d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 813.300856] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 813.301187] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 8.348s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 813.301390] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.145s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 813.302966] env[62753]: INFO nova.compute.claims [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 813.403987] env[62753]: DEBUG oslo_concurrency.lockutils [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 814.497155] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3ddf8f02-b008-43c8-ba1c-830118b1d4ab {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.505252] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89fd9d9b-64f5-45a3-b59b-c6c8558ad056 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.534257] env[62753]: INFO nova.compute.manager [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Rebuilding instance [ 814.536871] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d63814c8-824c-4050-ac0e-d4e804ae5477 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.543785] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-895c750c-9548-4e3c-9095-369654b4ace7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 814.558702] env[62753]: DEBUG nova.compute.provider_tree [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 814.578150] env[62753]: DEBUG nova.compute.manager [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 814.579124] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30c25da3-b5ee-4d03-96a2-97f066842a0c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.061906] env[62753]: DEBUG nova.scheduler.client.report [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 815.089986] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 815.090322] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f843ce9f-fbe0-41cc-9417-2e57f9fcb9fc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.098075] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Waiting for the task: (returnval){ [ 815.098075] env[62753]: value = "task-1332149" [ 815.098075] env[62753]: _type = "Task" [ 815.098075] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.107485] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': task-1332149, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 815.567287] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.266s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 815.568408] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.804s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 815.607636] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': task-1332149, 'name': PowerOffVM_Task, 'duration_secs': 0.129071} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 815.608387] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 815.608603] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 815.609345] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e10e2c0-07ad-4747-a90d-6e48f0179dd8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.616039] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 815.616243] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e081c91a-8763-45f1-9a3d-9bcfc14cac96 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.636619] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 815.636802] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 815.636972] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Deleting the datastore file [datastore1] 41476adb-95e5-4617-b178-9c981d13c43d {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 815.637202] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-10af8d83-a440-47ac-aa96-4ad192517103 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 815.642785] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Waiting for the task: (returnval){ [ 815.642785] env[62753]: value = "task-1332151" [ 815.642785] env[62753]: _type = "Task" [ 815.642785] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 815.650062] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': task-1332151, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 816.072481] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Acquiring lock "00d84c07-dab4-4d1c-9150-9b21003877e1" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 816.072734] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Lock "00d84c07-dab4-4d1c-9150-9b21003877e1" acquired by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 816.152370] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': task-1332151, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094291} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 816.154762] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 816.154951] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 816.155142] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 816.287606] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-450f8184-a4f6-4df5-9753-866380169a03 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.295429] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-823f47a8-7455-4c12-8df7-b1a9883e95a9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.325149] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5bef15a9-dec2-4155-9672-96ffad6f4ae7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.331974] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf70b277-c871-4f60-964f-4a3b5c271917 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 816.346468] env[62753]: DEBUG nova.compute.provider_tree [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 816.577746] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Lock "00d84c07-dab4-4d1c-9150-9b21003877e1" "released" by "nova.compute.manager.ComputeManager._validate_instance_group_policy.._do_validation" :: held 0.505s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 816.578349] env[62753]: DEBUG nova.compute.manager [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 816.849996] env[62753]: DEBUG nova.scheduler.client.report [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 817.083567] env[62753]: DEBUG nova.compute.utils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 817.084926] env[62753]: DEBUG nova.compute.manager [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 817.085146] env[62753]: DEBUG nova.network.neutron [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 817.136913] env[62753]: DEBUG nova.policy [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '3b083222441d48e7ad2241ff8d63af06', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2855858a3471457481fa2bf13af2b133', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 817.189809] env[62753]: DEBUG nova.virt.hardware [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 817.190089] env[62753]: DEBUG nova.virt.hardware [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 817.190200] env[62753]: DEBUG nova.virt.hardware [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 817.190437] env[62753]: DEBUG nova.virt.hardware [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 817.190594] env[62753]: DEBUG nova.virt.hardware [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 817.190742] env[62753]: DEBUG nova.virt.hardware [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 817.190977] env[62753]: DEBUG nova.virt.hardware [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 817.191243] env[62753]: DEBUG nova.virt.hardware [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 817.191299] env[62753]: DEBUG nova.virt.hardware [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 817.191444] env[62753]: DEBUG nova.virt.hardware [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 817.191616] env[62753]: DEBUG nova.virt.hardware [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 817.192610] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc460ced-b78d-4dc8-b779-f9e311f68d1e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.200619] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e641bf5-ca4e-4a54-a04d-90b911a66b05 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.214908] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Instance VIF info [] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 817.220541] env[62753]: DEBUG oslo.service.loopingcall [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 817.220813] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 817.221038] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d9e77ba0-f320-4408-a7a3-d41cf0f5e172 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.241133] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 817.241133] env[62753]: value = "task-1332153" [ 817.241133] env[62753]: _type = "Task" [ 817.241133] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.248931] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332153, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.355627] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.787s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 817.356764] env[62753]: ERROR nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 53d194f5-a7ce-4eb0-99da-e89986b6fc73, please check neutron logs for more information. [ 817.356764] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Traceback (most recent call last): [ 817.356764] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 817.356764] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] self.driver.spawn(context, instance, image_meta, [ 817.356764] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 817.356764] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] self._vmops.spawn(context, instance, image_meta, injected_files, [ 817.356764] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 817.356764] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] vm_ref = self.build_virtual_machine(instance, [ 817.356764] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 817.356764] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] vif_infos = vmwarevif.get_vif_info(self._session, [ 817.356764] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 817.357169] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] for vif in network_info: [ 817.357169] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 817.357169] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] return self._sync_wrapper(fn, *args, **kwargs) [ 817.357169] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 817.357169] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] self.wait() [ 817.357169] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 817.357169] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] self[:] = self._gt.wait() [ 817.357169] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 817.357169] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] return self._exit_event.wait() [ 817.357169] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 817.357169] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] current.throw(*self._exc) [ 817.357169] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 817.357169] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] result = function(*args, **kwargs) [ 817.357583] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 817.357583] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] return func(*args, **kwargs) [ 817.357583] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 817.357583] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] raise e [ 817.357583] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 817.357583] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] nwinfo = self.network_api.allocate_for_instance( [ 817.357583] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 817.357583] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] created_port_ids = self._update_ports_for_instance( [ 817.357583] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 817.357583] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] with excutils.save_and_reraise_exception(): [ 817.357583] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 817.357583] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] self.force_reraise() [ 817.357583] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 817.358008] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] raise self.value [ 817.358008] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 817.358008] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] updated_port = self._update_port( [ 817.358008] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 817.358008] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] _ensure_no_port_binding_failure(port) [ 817.358008] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 817.358008] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] raise exception.PortBindingFailed(port_id=port['id']) [ 817.358008] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] nova.exception.PortBindingFailed: Binding failed for port 53d194f5-a7ce-4eb0-99da-e89986b6fc73, please check neutron logs for more information. [ 817.358008] env[62753]: ERROR nova.compute.manager [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] [ 817.358008] env[62753]: DEBUG nova.compute.utils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Binding failed for port 53d194f5-a7ce-4eb0-99da-e89986b6fc73, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 817.359589] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.034s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 817.362018] env[62753]: INFO nova.compute.claims [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 817.366484] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Build of instance d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295 was re-scheduled: Binding failed for port 53d194f5-a7ce-4eb0-99da-e89986b6fc73, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 817.367226] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 817.367600] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquiring lock "refresh_cache-d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.367997] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Acquired lock "refresh_cache-d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.368155] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 817.424677] env[62753]: DEBUG nova.network.neutron [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Successfully created port: 61585ec2-05bb-498f-b46f-b26cd54d0f0c {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 817.588179] env[62753]: DEBUG nova.compute.manager [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 817.752976] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332153, 'name': CreateVM_Task, 'duration_secs': 0.29254} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 817.753228] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 817.753682] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 817.753889] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 817.754278] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 817.754575] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-71b7f5ff-84ec-4adf-98fc-b84c72e48bdd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 817.759427] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Waiting for the task: (returnval){ [ 817.759427] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52318491-f0c3-869e-3949-b1206d189f2d" [ 817.759427] env[62753]: _type = "Task" [ 817.759427] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 817.770059] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52318491-f0c3-869e-3949-b1206d189f2d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 817.896560] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 817.988815] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.135758] env[62753]: DEBUG nova.compute.manager [req-3df02b19-3c06-4941-aa2d-8fbdf6c05c00 req-f8b70e02-a96d-439e-94fd-939d6551f6be service nova] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Received event network-changed-61585ec2-05bb-498f-b46f-b26cd54d0f0c {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 818.135956] env[62753]: DEBUG nova.compute.manager [req-3df02b19-3c06-4941-aa2d-8fbdf6c05c00 req-f8b70e02-a96d-439e-94fd-939d6551f6be service nova] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Refreshing instance network info cache due to event network-changed-61585ec2-05bb-498f-b46f-b26cd54d0f0c. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 818.136199] env[62753]: DEBUG oslo_concurrency.lockutils [req-3df02b19-3c06-4941-aa2d-8fbdf6c05c00 req-f8b70e02-a96d-439e-94fd-939d6551f6be service nova] Acquiring lock "refresh_cache-07c4e02b-a8ca-41bb-82ec-38a81b9f1d07" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.136353] env[62753]: DEBUG oslo_concurrency.lockutils [req-3df02b19-3c06-4941-aa2d-8fbdf6c05c00 req-f8b70e02-a96d-439e-94fd-939d6551f6be service nova] Acquired lock "refresh_cache-07c4e02b-a8ca-41bb-82ec-38a81b9f1d07" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.136651] env[62753]: DEBUG nova.network.neutron [req-3df02b19-3c06-4941-aa2d-8fbdf6c05c00 req-f8b70e02-a96d-439e-94fd-939d6551f6be service nova] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Refreshing network info cache for port 61585ec2-05bb-498f-b46f-b26cd54d0f0c {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 818.270716] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52318491-f0c3-869e-3949-b1206d189f2d, 'name': SearchDatastore_Task, 'duration_secs': 0.00995} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.271021] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.271291] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 818.271564] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.271748] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 818.271958] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 818.272277] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-36f399a8-8945-42f1-ab84-fdafbf878d39 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.280367] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 818.280562] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 818.281257] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-743bf1b0-e9dd-4f35-9d76-d96a0c775fad {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.287091] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Waiting for the task: (returnval){ [ 818.287091] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ca4b36-9688-5fc0-82d1-aa5cb0b08161" [ 818.287091] env[62753]: _type = "Task" [ 818.287091] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.294861] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ca4b36-9688-5fc0-82d1-aa5cb0b08161, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 818.332221] env[62753]: ERROR nova.compute.manager [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 61585ec2-05bb-498f-b46f-b26cd54d0f0c, please check neutron logs for more information. [ 818.332221] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 818.332221] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 818.332221] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 818.332221] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 818.332221] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 818.332221] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 818.332221] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 818.332221] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 818.332221] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 818.332221] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 818.332221] env[62753]: ERROR nova.compute.manager raise self.value [ 818.332221] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 818.332221] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 818.332221] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 818.332221] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 818.333185] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 818.333185] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 818.333185] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 61585ec2-05bb-498f-b46f-b26cd54d0f0c, please check neutron logs for more information. [ 818.333185] env[62753]: ERROR nova.compute.manager [ 818.333185] env[62753]: Traceback (most recent call last): [ 818.333185] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 818.333185] env[62753]: listener.cb(fileno) [ 818.333185] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 818.333185] env[62753]: result = function(*args, **kwargs) [ 818.333185] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 818.333185] env[62753]: return func(*args, **kwargs) [ 818.333185] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 818.333185] env[62753]: raise e [ 818.333185] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 818.333185] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 818.333185] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 818.333185] env[62753]: created_port_ids = self._update_ports_for_instance( [ 818.333185] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 818.333185] env[62753]: with excutils.save_and_reraise_exception(): [ 818.333185] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 818.333185] env[62753]: self.force_reraise() [ 818.333185] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 818.333185] env[62753]: raise self.value [ 818.333185] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 818.333185] env[62753]: updated_port = self._update_port( [ 818.333185] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 818.333185] env[62753]: _ensure_no_port_binding_failure(port) [ 818.333185] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 818.333185] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 818.334730] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 61585ec2-05bb-498f-b46f-b26cd54d0f0c, please check neutron logs for more information. [ 818.334730] env[62753]: Removing descriptor: 18 [ 818.491988] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Releasing lock "refresh_cache-d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 818.492288] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 818.492469] env[62753]: DEBUG nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 818.492638] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 818.507515] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.597915] env[62753]: DEBUG nova.compute.manager [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 818.621695] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-921b4a8c-f665-434a-8db2-22c0d64f82a7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.632473] env[62753]: DEBUG nova.virt.hardware [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 818.632704] env[62753]: DEBUG nova.virt.hardware [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 818.632862] env[62753]: DEBUG nova.virt.hardware [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 818.633087] env[62753]: DEBUG nova.virt.hardware [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 818.633375] env[62753]: DEBUG nova.virt.hardware [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 818.633497] env[62753]: DEBUG nova.virt.hardware [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 818.633708] env[62753]: DEBUG nova.virt.hardware [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 818.633865] env[62753]: DEBUG nova.virt.hardware [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 818.634039] env[62753]: DEBUG nova.virt.hardware [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 818.634202] env[62753]: DEBUG nova.virt.hardware [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 818.634370] env[62753]: DEBUG nova.virt.hardware [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 818.635382] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84408604-8ef4-4581-996f-a6dad10a867d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.641768] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c3e22b9-f114-4470-98cf-c8b63c868e1d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.679420] env[62753]: DEBUG nova.network.neutron [req-3df02b19-3c06-4941-aa2d-8fbdf6c05c00 req-f8b70e02-a96d-439e-94fd-939d6551f6be service nova] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 818.682193] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efaf159d-f0c6-4cd3-a510-31525f25612e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.686404] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4c23fe8-65d4-49c9-a5c4-f114dae47628 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.706023] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c22acd1-e8f4-43e5-ac21-2283a2aef545 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.707598] env[62753]: ERROR nova.compute.manager [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 61585ec2-05bb-498f-b46f-b26cd54d0f0c, please check neutron logs for more information. [ 818.707598] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Traceback (most recent call last): [ 818.707598] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 818.707598] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] yield resources [ 818.707598] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 818.707598] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] self.driver.spawn(context, instance, image_meta, [ 818.707598] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 818.707598] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] self._vmops.spawn(context, instance, image_meta, injected_files, [ 818.707598] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 818.707598] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] vm_ref = self.build_virtual_machine(instance, [ 818.707598] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 818.707983] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] vif_infos = vmwarevif.get_vif_info(self._session, [ 818.707983] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 818.707983] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] for vif in network_info: [ 818.707983] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 818.707983] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] return self._sync_wrapper(fn, *args, **kwargs) [ 818.707983] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 818.707983] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] self.wait() [ 818.707983] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 818.707983] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] self[:] = self._gt.wait() [ 818.707983] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 818.707983] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] return self._exit_event.wait() [ 818.707983] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 818.707983] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] current.throw(*self._exc) [ 818.708363] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 818.708363] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] result = function(*args, **kwargs) [ 818.708363] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 818.708363] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] return func(*args, **kwargs) [ 818.708363] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 818.708363] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] raise e [ 818.708363] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 818.708363] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] nwinfo = self.network_api.allocate_for_instance( [ 818.708363] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 818.708363] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] created_port_ids = self._update_ports_for_instance( [ 818.708363] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 818.708363] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] with excutils.save_and_reraise_exception(): [ 818.708363] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 818.708876] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] self.force_reraise() [ 818.708876] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 818.708876] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] raise self.value [ 818.708876] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 818.708876] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] updated_port = self._update_port( [ 818.708876] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 818.708876] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] _ensure_no_port_binding_failure(port) [ 818.708876] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 818.708876] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] raise exception.PortBindingFailed(port_id=port['id']) [ 818.708876] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] nova.exception.PortBindingFailed: Binding failed for port 61585ec2-05bb-498f-b46f-b26cd54d0f0c, please check neutron logs for more information. [ 818.708876] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] [ 818.708876] env[62753]: INFO nova.compute.manager [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Terminating instance [ 818.711762] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Acquiring lock "refresh_cache-07c4e02b-a8ca-41bb-82ec-38a81b9f1d07" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 818.720394] env[62753]: DEBUG nova.compute.provider_tree [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 818.770099] env[62753]: DEBUG nova.network.neutron [req-3df02b19-3c06-4941-aa2d-8fbdf6c05c00 req-f8b70e02-a96d-439e-94fd-939d6551f6be service nova] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 818.797735] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ca4b36-9688-5fc0-82d1-aa5cb0b08161, 'name': SearchDatastore_Task, 'duration_secs': 0.008618} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 818.798434] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-efc92125-1c39-4fdd-9061-cf31ada37479 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 818.803481] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Waiting for the task: (returnval){ [ 818.803481] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52efaf1d-c2fe-5ae4-64db-dd901419eef3" [ 818.803481] env[62753]: _type = "Task" [ 818.803481] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 818.811041] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52efaf1d-c2fe-5ae4-64db-dd901419eef3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.012364] env[62753]: DEBUG nova.network.neutron [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 819.223323] env[62753]: DEBUG nova.scheduler.client.report [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 819.272339] env[62753]: DEBUG oslo_concurrency.lockutils [req-3df02b19-3c06-4941-aa2d-8fbdf6c05c00 req-f8b70e02-a96d-439e-94fd-939d6551f6be service nova] Releasing lock "refresh_cache-07c4e02b-a8ca-41bb-82ec-38a81b9f1d07" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.272736] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Acquired lock "refresh_cache-07c4e02b-a8ca-41bb-82ec-38a81b9f1d07" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 819.273670] env[62753]: DEBUG nova.network.neutron [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 819.313290] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52efaf1d-c2fe-5ae4-64db-dd901419eef3, 'name': SearchDatastore_Task, 'duration_secs': 0.008581} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.313541] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 819.313784] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 41476adb-95e5-4617-b178-9c981d13c43d/41476adb-95e5-4617-b178-9c981d13c43d.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 819.314040] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3f7b924b-2618-4aad-acca-06577e8360b2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.320673] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Waiting for the task: (returnval){ [ 819.320673] env[62753]: value = "task-1332154" [ 819.320673] env[62753]: _type = "Task" [ 819.320673] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.327990] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': task-1332154, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.515467] env[62753]: INFO nova.compute.manager [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] [instance: d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295] Took 1.02 seconds to deallocate network for instance. [ 819.728928] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.369s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 819.729662] env[62753]: DEBUG nova.compute.manager [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 819.732753] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 24.854s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 819.791674] env[62753]: DEBUG nova.network.neutron [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 819.833097] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': task-1332154, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507212} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 819.835287] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 41476adb-95e5-4617-b178-9c981d13c43d/41476adb-95e5-4617-b178-9c981d13c43d.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 819.835563] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 819.835823] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6b1006d4-b6ec-4c80-a830-a2201aae69af {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 819.842095] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Waiting for the task: (returnval){ [ 819.842095] env[62753]: value = "task-1332155" [ 819.842095] env[62753]: _type = "Task" [ 819.842095] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 819.849282] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': task-1332155, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 819.871882] env[62753]: DEBUG nova.network.neutron [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 820.189543] env[62753]: DEBUG nova.compute.manager [req-f18fd8bb-9884-4e1e-8a74-1d61fb422e9b req-984b8985-8959-4413-bdfc-aa8908c9b82f service nova] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Received event network-vif-deleted-61585ec2-05bb-498f-b46f-b26cd54d0f0c {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 820.237777] env[62753]: DEBUG nova.compute.utils [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 820.242284] env[62753]: DEBUG nova.compute.manager [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 820.243979] env[62753]: DEBUG nova.network.neutron [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 820.293859] env[62753]: DEBUG nova.policy [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4dbc63095b0464fa0d2de86e3cf170e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5636da907ab343f9b42dbbd903d32283', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 820.352775] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': task-1332155, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075628} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.352907] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 820.353704] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d702790-63e8-45fd-877b-b032136eec02 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.375246] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Reconfiguring VM instance instance-00000030 to attach disk [datastore1] 41476adb-95e5-4617-b178-9c981d13c43d/41476adb-95e5-4617-b178-9c981d13c43d.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 820.377853] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Releasing lock "refresh_cache-07c4e02b-a8ca-41bb-82ec-38a81b9f1d07" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 820.378238] env[62753]: DEBUG nova.compute.manager [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 820.378426] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 820.378663] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-deea21ea-c338-4d14-a8d6-36d3f0f10341 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.393056] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5167acb0-2cd9-49d2-81bb-20d4d0e8a568 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.403161] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a879a821-0e65-4be1-8338-da217a149ab8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.413267] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Waiting for the task: (returnval){ [ 820.413267] env[62753]: value = "task-1332156" [ 820.413267] env[62753]: _type = "Task" [ 820.413267] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.426515] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': task-1332156, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.429994] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07 could not be found. [ 820.430259] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 820.430498] env[62753]: INFO nova.compute.manager [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Took 0.05 seconds to destroy the instance on the hypervisor. [ 820.430803] env[62753]: DEBUG oslo.service.loopingcall [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 820.433416] env[62753]: DEBUG nova.compute.manager [-] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 820.433515] env[62753]: DEBUG nova.network.neutron [-] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 820.449845] env[62753]: DEBUG nova.network.neutron [-] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 820.534192] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f2d0642-7a00-44bc-a891-22eb9220c55f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.541927] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc9d15f0-86be-4db7-99a7-db1c199c3563 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.547509] env[62753]: INFO nova.scheduler.client.report [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Deleted allocations for instance d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295 [ 820.553061] env[62753]: DEBUG nova.network.neutron [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Successfully created port: a0fe6d5a-8a68-4ab9-8d30-b29003f609be {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 820.592582] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27a8e456-9bed-471b-9568-dc1a605f6df2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.603160] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43fa9165-d1eb-4a22-9782-f2d8206581cb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.617752] env[62753]: DEBUG nova.compute.provider_tree [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 820.726123] env[62753]: DEBUG oslo_concurrency.lockutils [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Acquiring lock "c5dae9a3-052b-4f4c-ac79-84aeaded457a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 820.726381] env[62753]: DEBUG oslo_concurrency.lockutils [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Lock "c5dae9a3-052b-4f4c-ac79-84aeaded457a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 820.743195] env[62753]: DEBUG nova.compute.manager [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 820.928523] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': task-1332156, 'name': ReconfigVM_Task, 'duration_secs': 0.286491} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 820.928801] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Reconfigured VM instance instance-00000030 to attach disk [datastore1] 41476adb-95e5-4617-b178-9c981d13c43d/41476adb-95e5-4617-b178-9c981d13c43d.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 820.929701] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-684ad234-3528-441d-b52e-347dd5e7abad {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 820.937493] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Waiting for the task: (returnval){ [ 820.937493] env[62753]: value = "task-1332157" [ 820.937493] env[62753]: _type = "Task" [ 820.937493] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 820.945446] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': task-1332157, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 820.951830] env[62753]: DEBUG nova.network.neutron [-] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 821.009827] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 821.009984] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Cleaning up deleted instances {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 821.057500] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d2329e9e-dc01-47a3-b974-fd972a8bf137 tempest-MultipleCreateTestJSON-867991426 tempest-MultipleCreateTestJSON-867991426-project-member] Lock "d1cbcc94-8ff0-4b8b-a98c-dbf46eaa4295" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 157.226s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.123018] env[62753]: DEBUG nova.scheduler.client.report [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 821.417174] env[62753]: ERROR nova.compute.manager [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port a0fe6d5a-8a68-4ab9-8d30-b29003f609be, please check neutron logs for more information. [ 821.417174] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 821.417174] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 821.417174] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 821.417174] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 821.417174] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 821.417174] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 821.417174] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 821.417174] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 821.417174] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 821.417174] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 821.417174] env[62753]: ERROR nova.compute.manager raise self.value [ 821.417174] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 821.417174] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 821.417174] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 821.417174] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 821.417816] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 821.417816] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 821.417816] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port a0fe6d5a-8a68-4ab9-8d30-b29003f609be, please check neutron logs for more information. [ 821.417816] env[62753]: ERROR nova.compute.manager [ 821.417816] env[62753]: Traceback (most recent call last): [ 821.417816] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 821.417816] env[62753]: listener.cb(fileno) [ 821.417816] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 821.417816] env[62753]: result = function(*args, **kwargs) [ 821.417816] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 821.417816] env[62753]: return func(*args, **kwargs) [ 821.417816] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 821.417816] env[62753]: raise e [ 821.417816] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 821.417816] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 821.417816] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 821.417816] env[62753]: created_port_ids = self._update_ports_for_instance( [ 821.417816] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 821.417816] env[62753]: with excutils.save_and_reraise_exception(): [ 821.417816] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 821.417816] env[62753]: self.force_reraise() [ 821.417816] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 821.417816] env[62753]: raise self.value [ 821.417816] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 821.417816] env[62753]: updated_port = self._update_port( [ 821.417816] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 821.417816] env[62753]: _ensure_no_port_binding_failure(port) [ 821.417816] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 821.417816] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 821.419062] env[62753]: nova.exception.PortBindingFailed: Binding failed for port a0fe6d5a-8a68-4ab9-8d30-b29003f609be, please check neutron logs for more information. [ 821.419062] env[62753]: Removing descriptor: 18 [ 821.447036] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': task-1332157, 'name': Rename_Task, 'duration_secs': 0.192938} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.447329] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 821.447570] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-af49835c-5283-4772-bca7-64a7a208af27 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.454707] env[62753]: INFO nova.compute.manager [-] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Took 1.02 seconds to deallocate network for instance. [ 821.455027] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Waiting for the task: (returnval){ [ 821.455027] env[62753]: value = "task-1332158" [ 821.455027] env[62753]: _type = "Task" [ 821.455027] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 821.459574] env[62753]: DEBUG nova.compute.claims [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 821.459740] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 821.464936] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': task-1332158, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 821.513864] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] There are 3 instances to clean {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 821.514183] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 8664ed5c-ecb9-4795-8499-31198cfd0450] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 821.560042] env[62753]: DEBUG nova.compute.manager [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 821.625812] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.893s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 821.626494] env[62753]: ERROR nova.compute.manager [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 257c7620-1533-42cc-b45b-b6615400c2d3, please check neutron logs for more information. [ 821.626494] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Traceback (most recent call last): [ 821.626494] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 821.626494] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] self.driver.spawn(context, instance, image_meta, [ 821.626494] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 821.626494] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] self._vmops.spawn(context, instance, image_meta, injected_files, [ 821.626494] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 821.626494] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] vm_ref = self.build_virtual_machine(instance, [ 821.626494] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 821.626494] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] vif_infos = vmwarevif.get_vif_info(self._session, [ 821.626494] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 821.626827] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] for vif in network_info: [ 821.626827] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 821.626827] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] return self._sync_wrapper(fn, *args, **kwargs) [ 821.626827] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 821.626827] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] self.wait() [ 821.626827] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 821.626827] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] self[:] = self._gt.wait() [ 821.626827] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 821.626827] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] return self._exit_event.wait() [ 821.626827] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 821.626827] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] current.throw(*self._exc) [ 821.626827] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 821.626827] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] result = function(*args, **kwargs) [ 821.627271] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 821.627271] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] return func(*args, **kwargs) [ 821.627271] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 821.627271] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] raise e [ 821.627271] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 821.627271] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] nwinfo = self.network_api.allocate_for_instance( [ 821.627271] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 821.627271] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] created_port_ids = self._update_ports_for_instance( [ 821.627271] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 821.627271] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] with excutils.save_and_reraise_exception(): [ 821.627271] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 821.627271] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] self.force_reraise() [ 821.627271] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 821.627889] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] raise self.value [ 821.627889] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 821.627889] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] updated_port = self._update_port( [ 821.627889] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 821.627889] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] _ensure_no_port_binding_failure(port) [ 821.627889] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 821.627889] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] raise exception.PortBindingFailed(port_id=port['id']) [ 821.627889] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] nova.exception.PortBindingFailed: Binding failed for port 257c7620-1533-42cc-b45b-b6615400c2d3, please check neutron logs for more information. [ 821.627889] env[62753]: ERROR nova.compute.manager [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] [ 821.627889] env[62753]: DEBUG nova.compute.utils [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Binding failed for port 257c7620-1533-42cc-b45b-b6615400c2d3, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 821.628438] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.192s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 821.629963] env[62753]: INFO nova.compute.claims [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 821.633058] env[62753]: DEBUG nova.compute.manager [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Build of instance e0c4a1d2-6a85-4826-88fa-d21556d744ac was re-scheduled: Binding failed for port 257c7620-1533-42cc-b45b-b6615400c2d3, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 821.633534] env[62753]: DEBUG nova.compute.manager [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 821.633762] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Acquiring lock "refresh_cache-e0c4a1d2-6a85-4826-88fa-d21556d744ac" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.633907] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Acquired lock "refresh_cache-e0c4a1d2-6a85-4826-88fa-d21556d744ac" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.634077] env[62753]: DEBUG nova.network.neutron [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 821.754599] env[62753]: DEBUG nova.compute.manager [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 821.781468] env[62753]: DEBUG nova.virt.hardware [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 821.781732] env[62753]: DEBUG nova.virt.hardware [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 821.781944] env[62753]: DEBUG nova.virt.hardware [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 821.782201] env[62753]: DEBUG nova.virt.hardware [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 821.782417] env[62753]: DEBUG nova.virt.hardware [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 821.782615] env[62753]: DEBUG nova.virt.hardware [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 821.782869] env[62753]: DEBUG nova.virt.hardware [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 821.783067] env[62753]: DEBUG nova.virt.hardware [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 821.783258] env[62753]: DEBUG nova.virt.hardware [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 821.783456] env[62753]: DEBUG nova.virt.hardware [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 821.783883] env[62753]: DEBUG nova.virt.hardware [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 821.784864] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4966b904-29c2-4fe6-91d9-d351e8732fc9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.796146] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf190b41-e2d9-44ed-94d9-333054143ebd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 821.812088] env[62753]: ERROR nova.compute.manager [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port a0fe6d5a-8a68-4ab9-8d30-b29003f609be, please check neutron logs for more information. [ 821.812088] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Traceback (most recent call last): [ 821.812088] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 821.812088] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] yield resources [ 821.812088] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 821.812088] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] self.driver.spawn(context, instance, image_meta, [ 821.812088] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 821.812088] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 821.812088] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 821.812088] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] vm_ref = self.build_virtual_machine(instance, [ 821.812088] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 821.812629] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 821.812629] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 821.812629] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] for vif in network_info: [ 821.812629] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 821.812629] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] return self._sync_wrapper(fn, *args, **kwargs) [ 821.812629] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 821.812629] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] self.wait() [ 821.812629] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 821.812629] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] self[:] = self._gt.wait() [ 821.812629] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 821.812629] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] return self._exit_event.wait() [ 821.812629] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 821.812629] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] current.throw(*self._exc) [ 821.813065] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 821.813065] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] result = function(*args, **kwargs) [ 821.813065] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 821.813065] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] return func(*args, **kwargs) [ 821.813065] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 821.813065] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] raise e [ 821.813065] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 821.813065] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] nwinfo = self.network_api.allocate_for_instance( [ 821.813065] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 821.813065] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] created_port_ids = self._update_ports_for_instance( [ 821.813065] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 821.813065] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] with excutils.save_and_reraise_exception(): [ 821.813065] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 821.813555] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] self.force_reraise() [ 821.813555] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 821.813555] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] raise self.value [ 821.813555] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 821.813555] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] updated_port = self._update_port( [ 821.813555] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 821.813555] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] _ensure_no_port_binding_failure(port) [ 821.813555] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 821.813555] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] raise exception.PortBindingFailed(port_id=port['id']) [ 821.813555] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] nova.exception.PortBindingFailed: Binding failed for port a0fe6d5a-8a68-4ab9-8d30-b29003f609be, please check neutron logs for more information. [ 821.813555] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] [ 821.813555] env[62753]: INFO nova.compute.manager [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Terminating instance [ 821.814837] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "refresh_cache-67f5d96c-b870-442a-a0f3-a70533c045ea" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 821.815011] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "refresh_cache-67f5d96c-b870-442a-a0f3-a70533c045ea" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 821.815280] env[62753]: DEBUG nova.network.neutron [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 821.965896] env[62753]: DEBUG oslo_vmware.api [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Task: {'id': task-1332158, 'name': PowerOnVM_Task, 'duration_secs': 0.429664} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 821.966189] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 821.966389] env[62753]: DEBUG nova.compute.manager [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 821.967218] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e00ccd4-9d00-4df6-a04f-840ee152843d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.018647] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: a8ceac07-6631-4ac6-b62a-7f3eac788f2b] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 822.083044] env[62753]: DEBUG oslo_concurrency.lockutils [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.154120] env[62753]: DEBUG nova.network.neutron [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.217725] env[62753]: DEBUG nova.compute.manager [req-84501eef-2db7-428f-b86e-09549341f362 req-bc98b522-5301-426e-ab39-960a5b47b2e6 service nova] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Received event network-changed-a0fe6d5a-8a68-4ab9-8d30-b29003f609be {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 822.217904] env[62753]: DEBUG nova.compute.manager [req-84501eef-2db7-428f-b86e-09549341f362 req-bc98b522-5301-426e-ab39-960a5b47b2e6 service nova] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Refreshing instance network info cache due to event network-changed-a0fe6d5a-8a68-4ab9-8d30-b29003f609be. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 822.218110] env[62753]: DEBUG oslo_concurrency.lockutils [req-84501eef-2db7-428f-b86e-09549341f362 req-bc98b522-5301-426e-ab39-960a5b47b2e6 service nova] Acquiring lock "refresh_cache-67f5d96c-b870-442a-a0f3-a70533c045ea" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 822.281021] env[62753]: DEBUG nova.network.neutron [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.457771] env[62753]: DEBUG nova.network.neutron [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.484789] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 822.521470] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 29dfb4db-c9ae-4a5c-8574-b27b13cdc83c] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 822.578766] env[62753]: DEBUG nova.network.neutron [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 822.785365] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Releasing lock "refresh_cache-e0c4a1d2-6a85-4826-88fa-d21556d744ac" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 822.785365] env[62753]: DEBUG nova.compute.manager [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 822.785365] env[62753]: DEBUG nova.compute.manager [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 822.785365] env[62753]: DEBUG nova.network.neutron [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 822.805907] env[62753]: DEBUG nova.network.neutron [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 822.893299] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd010044-3b73-4f0a-85fe-21704f966e14 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.901018] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-269f9365-8cbd-460e-ab7a-7ed39e9a8c5d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.929871] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9eed338-b930-4862-b674-66bd1b788d59 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.937403] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e0673a03-2b55-42df-ad95-ea65cf4e3516 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 822.950581] env[62753]: DEBUG nova.compute.provider_tree [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 823.025143] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 823.025383] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Cleaning up deleted instances with incomplete migration {{(pid=62753) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 823.083568] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "refresh_cache-67f5d96c-b870-442a-a0f3-a70533c045ea" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 823.083568] env[62753]: DEBUG nova.compute.manager [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 823.083568] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 823.083568] env[62753]: DEBUG oslo_concurrency.lockutils [req-84501eef-2db7-428f-b86e-09549341f362 req-bc98b522-5301-426e-ab39-960a5b47b2e6 service nova] Acquired lock "refresh_cache-67f5d96c-b870-442a-a0f3-a70533c045ea" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.083568] env[62753]: DEBUG nova.network.neutron [req-84501eef-2db7-428f-b86e-09549341f362 req-bc98b522-5301-426e-ab39-960a5b47b2e6 service nova] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Refreshing network info cache for port a0fe6d5a-8a68-4ab9-8d30-b29003f609be {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 823.083989] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-f5e0356c-f72b-484f-809f-1998d312cec3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.099698] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8607ac0e-d589-4036-8a7a-3f13b52d5a48 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 823.130389] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 67f5d96c-b870-442a-a0f3-a70533c045ea could not be found. [ 823.130766] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 823.130894] env[62753]: INFO nova.compute.manager [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Took 0.05 seconds to destroy the instance on the hypervisor. [ 823.131157] env[62753]: DEBUG oslo.service.loopingcall [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 823.131384] env[62753]: DEBUG nova.compute.manager [-] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 823.131478] env[62753]: DEBUG nova.network.neutron [-] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 823.151053] env[62753]: DEBUG nova.network.neutron [-] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.308686] env[62753]: DEBUG nova.network.neutron [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.455161] env[62753]: DEBUG nova.scheduler.client.report [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 823.528660] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 823.611280] env[62753]: DEBUG nova.network.neutron [req-84501eef-2db7-428f-b86e-09549341f362 req-bc98b522-5301-426e-ab39-960a5b47b2e6 service nova] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 823.642591] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquiring lock "41476adb-95e5-4617-b178-9c981d13c43d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.642852] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Lock "41476adb-95e5-4617-b178-9c981d13c43d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.643084] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquiring lock "41476adb-95e5-4617-b178-9c981d13c43d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 823.643409] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Lock "41476adb-95e5-4617-b178-9c981d13c43d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.643588] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Lock "41476adb-95e5-4617-b178-9c981d13c43d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.646425] env[62753]: INFO nova.compute.manager [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Terminating instance [ 823.648533] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquiring lock "refresh_cache-41476adb-95e5-4617-b178-9c981d13c43d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 823.648798] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquired lock "refresh_cache-41476adb-95e5-4617-b178-9c981d13c43d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 823.648924] env[62753]: DEBUG nova.network.neutron [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 823.653339] env[62753]: DEBUG nova.network.neutron [-] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.734230] env[62753]: DEBUG nova.network.neutron [req-84501eef-2db7-428f-b86e-09549341f362 req-bc98b522-5301-426e-ab39-960a5b47b2e6 service nova] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 823.811285] env[62753]: INFO nova.compute.manager [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] [instance: e0c4a1d2-6a85-4826-88fa-d21556d744ac] Took 1.03 seconds to deallocate network for instance. [ 823.959439] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.331s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 823.959961] env[62753]: DEBUG nova.compute.manager [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 823.963059] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 25.616s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 823.964516] env[62753]: INFO nova.compute.claims [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 824.158361] env[62753]: INFO nova.compute.manager [-] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Took 1.03 seconds to deallocate network for instance. [ 824.159327] env[62753]: DEBUG nova.compute.claims [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 824.159653] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 824.170709] env[62753]: DEBUG nova.network.neutron [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 824.224348] env[62753]: DEBUG nova.network.neutron [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 824.236398] env[62753]: DEBUG oslo_concurrency.lockutils [req-84501eef-2db7-428f-b86e-09549341f362 req-bc98b522-5301-426e-ab39-960a5b47b2e6 service nova] Releasing lock "refresh_cache-67f5d96c-b870-442a-a0f3-a70533c045ea" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.236651] env[62753]: DEBUG nova.compute.manager [req-84501eef-2db7-428f-b86e-09549341f362 req-bc98b522-5301-426e-ab39-960a5b47b2e6 service nova] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Received event network-vif-deleted-a0fe6d5a-8a68-4ab9-8d30-b29003f609be {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 824.468902] env[62753]: DEBUG nova.compute.utils [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 824.471642] env[62753]: DEBUG nova.compute.manager [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 824.471809] env[62753]: DEBUG nova.network.neutron [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 824.514038] env[62753]: DEBUG nova.policy [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'e8874498ef9f4c08820b76b7d424547f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b8f42babc4204d25979a2ed530963d9e', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 824.726809] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Releasing lock "refresh_cache-41476adb-95e5-4617-b178-9c981d13c43d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 824.727287] env[62753]: DEBUG nova.compute.manager [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 824.727488] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 824.728865] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-481a3f01-7044-4360-9fad-7cb4cbfb90be {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.738664] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 824.738955] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f4a593bd-db29-4beb-a34b-2fac8ebd4fc3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 824.745168] env[62753]: DEBUG oslo_vmware.api [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 824.745168] env[62753]: value = "task-1332159" [ 824.745168] env[62753]: _type = "Task" [ 824.745168] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 824.754022] env[62753]: DEBUG oslo_vmware.api [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332159, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 824.788624] env[62753]: DEBUG nova.network.neutron [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Successfully created port: 4aef7cff-9e33-4256-9f64-2960ecceb14b {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 824.847069] env[62753]: INFO nova.scheduler.client.report [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Deleted allocations for instance e0c4a1d2-6a85-4826-88fa-d21556d744ac [ 824.977461] env[62753]: DEBUG nova.compute.manager [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 825.032038] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.032417] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.032496] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.032579] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.032961] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.033130] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.033300] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.033436] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62753) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 825.033579] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 825.210751] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a04d3241-526b-43bb-b9ac-5c5199ae9089 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.218478] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1cc25ce-2f71-4d63-a2a5-583480d5b195 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.255825] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-345fb1b5-48b9-444e-8038-bdf19b3944a3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.263272] env[62753]: DEBUG oslo_vmware.api [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332159, 'name': PowerOffVM_Task, 'duration_secs': 0.21419} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.265452] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 825.265513] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 825.265767] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-59531aaf-14b5-47c4-9051-a125d10f6ed2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.268510] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f04cc59-387c-4b73-8e81-9027bff910f9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.284833] env[62753]: DEBUG nova.compute.provider_tree [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 825.293706] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 825.293904] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 825.294333] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Deleting the datastore file [datastore1] 41476adb-95e5-4617-b178-9c981d13c43d {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 825.294333] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-784e0b39-8410-450c-a632-ea1b4ca2f3b0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 825.300405] env[62753]: DEBUG oslo_vmware.api [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for the task: (returnval){ [ 825.300405] env[62753]: value = "task-1332161" [ 825.300405] env[62753]: _type = "Task" [ 825.300405] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 825.309383] env[62753]: DEBUG oslo_vmware.api [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332161, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 825.355905] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cfc10e60-0d92-44f8-a497-259407e541e3 tempest-InstanceActionsNegativeTestJSON-1352231441 tempest-InstanceActionsNegativeTestJSON-1352231441-project-member] Lock "e0c4a1d2-6a85-4826-88fa-d21556d744ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 145.026s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 825.515892] env[62753]: DEBUG nova.compute.manager [req-0c11c6da-b84d-4f04-8661-0a3228b54652 req-d2eb0560-07ed-4127-af83-2aebfb07af6f service nova] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Received event network-changed-4aef7cff-9e33-4256-9f64-2960ecceb14b {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 825.516086] env[62753]: DEBUG nova.compute.manager [req-0c11c6da-b84d-4f04-8661-0a3228b54652 req-d2eb0560-07ed-4127-af83-2aebfb07af6f service nova] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Refreshing instance network info cache due to event network-changed-4aef7cff-9e33-4256-9f64-2960ecceb14b. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 825.516308] env[62753]: DEBUG oslo_concurrency.lockutils [req-0c11c6da-b84d-4f04-8661-0a3228b54652 req-d2eb0560-07ed-4127-af83-2aebfb07af6f service nova] Acquiring lock "refresh_cache-92e86d6a-e455-4984-9fdb-d1cdf856e729" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 825.516443] env[62753]: DEBUG oslo_concurrency.lockutils [req-0c11c6da-b84d-4f04-8661-0a3228b54652 req-d2eb0560-07ed-4127-af83-2aebfb07af6f service nova] Acquired lock "refresh_cache-92e86d6a-e455-4984-9fdb-d1cdf856e729" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 825.517336] env[62753]: DEBUG nova.network.neutron [req-0c11c6da-b84d-4f04-8661-0a3228b54652 req-d2eb0560-07ed-4127-af83-2aebfb07af6f service nova] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Refreshing network info cache for port 4aef7cff-9e33-4256-9f64-2960ecceb14b {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 825.536529] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 825.641357] env[62753]: ERROR nova.compute.manager [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4aef7cff-9e33-4256-9f64-2960ecceb14b, please check neutron logs for more information. [ 825.641357] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 825.641357] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 825.641357] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 825.641357] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 825.641357] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 825.641357] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 825.641357] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 825.641357] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 825.641357] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 825.641357] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 825.641357] env[62753]: ERROR nova.compute.manager raise self.value [ 825.641357] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 825.641357] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 825.641357] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 825.641357] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 825.642101] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 825.642101] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 825.642101] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4aef7cff-9e33-4256-9f64-2960ecceb14b, please check neutron logs for more information. [ 825.642101] env[62753]: ERROR nova.compute.manager [ 825.642101] env[62753]: Traceback (most recent call last): [ 825.642101] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 825.642101] env[62753]: listener.cb(fileno) [ 825.642101] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 825.642101] env[62753]: result = function(*args, **kwargs) [ 825.642101] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 825.642101] env[62753]: return func(*args, **kwargs) [ 825.642101] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 825.642101] env[62753]: raise e [ 825.642101] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 825.642101] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 825.642101] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 825.642101] env[62753]: created_port_ids = self._update_ports_for_instance( [ 825.642101] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 825.642101] env[62753]: with excutils.save_and_reraise_exception(): [ 825.642101] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 825.642101] env[62753]: self.force_reraise() [ 825.642101] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 825.642101] env[62753]: raise self.value [ 825.642101] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 825.642101] env[62753]: updated_port = self._update_port( [ 825.642101] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 825.642101] env[62753]: _ensure_no_port_binding_failure(port) [ 825.642101] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 825.642101] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 825.642919] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 4aef7cff-9e33-4256-9f64-2960ecceb14b, please check neutron logs for more information. [ 825.642919] env[62753]: Removing descriptor: 18 [ 825.788568] env[62753]: DEBUG nova.scheduler.client.report [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 825.815092] env[62753]: DEBUG oslo_vmware.api [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Task: {'id': task-1332161, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.094661} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 825.815092] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 825.815092] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 825.815092] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 825.815092] env[62753]: INFO nova.compute.manager [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Took 1.09 seconds to destroy the instance on the hypervisor. [ 825.815494] env[62753]: DEBUG oslo.service.loopingcall [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 825.815494] env[62753]: DEBUG nova.compute.manager [-] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 825.816214] env[62753]: DEBUG nova.network.neutron [-] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 825.832865] env[62753]: DEBUG nova.network.neutron [-] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 825.860845] env[62753]: DEBUG nova.compute.manager [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 825.993795] env[62753]: DEBUG nova.compute.manager [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 826.019119] env[62753]: DEBUG nova.virt.hardware [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 826.019119] env[62753]: DEBUG nova.virt.hardware [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 826.019119] env[62753]: DEBUG nova.virt.hardware [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 826.019414] env[62753]: DEBUG nova.virt.hardware [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 826.019414] env[62753]: DEBUG nova.virt.hardware [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 826.019414] env[62753]: DEBUG nova.virt.hardware [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 826.019414] env[62753]: DEBUG nova.virt.hardware [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 826.019830] env[62753]: DEBUG nova.virt.hardware [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 826.020241] env[62753]: DEBUG nova.virt.hardware [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 826.020626] env[62753]: DEBUG nova.virt.hardware [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 826.021522] env[62753]: DEBUG nova.virt.hardware [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 826.022142] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-034c10e5-6c40-4fa1-9057-db210198d2a7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.033488] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b5058c9-6078-48f8-b692-4375788386d9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 826.048887] env[62753]: ERROR nova.compute.manager [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4aef7cff-9e33-4256-9f64-2960ecceb14b, please check neutron logs for more information. [ 826.048887] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Traceback (most recent call last): [ 826.048887] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 826.048887] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] yield resources [ 826.048887] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 826.048887] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] self.driver.spawn(context, instance, image_meta, [ 826.048887] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 826.048887] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] self._vmops.spawn(context, instance, image_meta, injected_files, [ 826.048887] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 826.048887] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] vm_ref = self.build_virtual_machine(instance, [ 826.048887] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 826.051139] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] vif_infos = vmwarevif.get_vif_info(self._session, [ 826.051139] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 826.051139] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] for vif in network_info: [ 826.051139] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 826.051139] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] return self._sync_wrapper(fn, *args, **kwargs) [ 826.051139] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 826.051139] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] self.wait() [ 826.051139] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 826.051139] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] self[:] = self._gt.wait() [ 826.051139] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 826.051139] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] return self._exit_event.wait() [ 826.051139] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 826.051139] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] current.throw(*self._exc) [ 826.051572] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 826.051572] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] result = function(*args, **kwargs) [ 826.051572] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 826.051572] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] return func(*args, **kwargs) [ 826.051572] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 826.051572] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] raise e [ 826.051572] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 826.051572] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] nwinfo = self.network_api.allocate_for_instance( [ 826.051572] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 826.051572] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] created_port_ids = self._update_ports_for_instance( [ 826.051572] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 826.051572] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] with excutils.save_and_reraise_exception(): [ 826.051572] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 826.051976] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] self.force_reraise() [ 826.051976] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 826.051976] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] raise self.value [ 826.051976] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 826.051976] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] updated_port = self._update_port( [ 826.051976] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 826.051976] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] _ensure_no_port_binding_failure(port) [ 826.051976] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 826.051976] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] raise exception.PortBindingFailed(port_id=port['id']) [ 826.051976] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] nova.exception.PortBindingFailed: Binding failed for port 4aef7cff-9e33-4256-9f64-2960ecceb14b, please check neutron logs for more information. [ 826.051976] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] [ 826.051976] env[62753]: INFO nova.compute.manager [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Terminating instance [ 826.055891] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Acquiring lock "refresh_cache-92e86d6a-e455-4984-9fdb-d1cdf856e729" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 826.055891] env[62753]: DEBUG nova.network.neutron [req-0c11c6da-b84d-4f04-8661-0a3228b54652 req-d2eb0560-07ed-4127-af83-2aebfb07af6f service nova] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 826.172868] env[62753]: DEBUG nova.network.neutron [req-0c11c6da-b84d-4f04-8661-0a3228b54652 req-d2eb0560-07ed-4127-af83-2aebfb07af6f service nova] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.296016] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.331s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 826.296016] env[62753]: DEBUG nova.compute.manager [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 826.304016] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 21.241s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 826.335694] env[62753]: DEBUG nova.network.neutron [-] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 826.387205] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 826.675263] env[62753]: DEBUG oslo_concurrency.lockutils [req-0c11c6da-b84d-4f04-8661-0a3228b54652 req-d2eb0560-07ed-4127-af83-2aebfb07af6f service nova] Releasing lock "refresh_cache-92e86d6a-e455-4984-9fdb-d1cdf856e729" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 826.675679] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Acquired lock "refresh_cache-92e86d6a-e455-4984-9fdb-d1cdf856e729" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 826.676794] env[62753]: DEBUG nova.network.neutron [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 826.815397] env[62753]: DEBUG nova.compute.utils [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 826.817914] env[62753]: DEBUG nova.compute.manager [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 826.817914] env[62753]: DEBUG nova.network.neutron [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 826.837732] env[62753]: INFO nova.compute.manager [-] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Took 1.02 seconds to deallocate network for instance. [ 826.891298] env[62753]: DEBUG nova.policy [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '613b397bebf9498596d07d27b9fa1fa6', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'b16df868b31c4f6597972c6f2efa9494', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 827.086771] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-022cac75-7ce1-4fd7-b0dd-335226ccbb6b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.095675] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c17a559-e503-421e-a2f6-68fad015c91b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.130755] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ea4c22b-7413-473b-982f-c4625aeb5ced {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.140605] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-247fb0e2-49a9-42b0-bff4-c60260c7ae81 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.154780] env[62753]: DEBUG nova.compute.provider_tree [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 827.205533] env[62753]: DEBUG nova.network.neutron [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 827.305123] env[62753]: DEBUG nova.network.neutron [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Successfully created port: 60dffaa7-7a7a-4d14-88e2-fc60b0015fd0 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 827.326564] env[62753]: DEBUG nova.compute.manager [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 827.346642] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 827.389427] env[62753]: DEBUG nova.network.neutron [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 827.550275] env[62753]: DEBUG nova.compute.manager [req-26fd8643-6f24-4a8a-a1c2-4780d2b06651 req-d8006803-a4b9-4dba-b1b2-1d9e093076a4 service nova] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Received event network-vif-deleted-4aef7cff-9e33-4256-9f64-2960ecceb14b {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 827.657935] env[62753]: DEBUG nova.scheduler.client.report [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 827.893039] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Releasing lock "refresh_cache-92e86d6a-e455-4984-9fdb-d1cdf856e729" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 827.893361] env[62753]: DEBUG nova.compute.manager [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 827.893570] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 827.893859] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-14613324-6cca-4789-bf4e-f1e8ec64bc39 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.906040] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9888c16c-314c-42d4-a1ca-f68fe1504420 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 827.929917] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 92e86d6a-e455-4984-9fdb-d1cdf856e729 could not be found. [ 827.930734] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 827.930734] env[62753]: INFO nova.compute.manager [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Took 0.04 seconds to destroy the instance on the hypervisor. [ 827.930734] env[62753]: DEBUG oslo.service.loopingcall [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 827.931565] env[62753]: DEBUG nova.compute.manager [-] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 827.931565] env[62753]: DEBUG nova.network.neutron [-] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 827.979352] env[62753]: DEBUG nova.network.neutron [-] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.043982] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "4d20e083-2959-453a-8875-47955bc02613" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 828.044233] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "4d20e083-2959-453a-8875-47955bc02613" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.164333] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.867s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 828.164943] env[62753]: ERROR nova.compute.manager [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 04dda318-4cf1-44d4-bf27-fe6dce719865, please check neutron logs for more information. [ 828.164943] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Traceback (most recent call last): [ 828.164943] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 828.164943] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] self.driver.spawn(context, instance, image_meta, [ 828.164943] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 828.164943] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] self._vmops.spawn(context, instance, image_meta, injected_files, [ 828.164943] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 828.164943] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] vm_ref = self.build_virtual_machine(instance, [ 828.164943] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 828.164943] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] vif_infos = vmwarevif.get_vif_info(self._session, [ 828.164943] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 828.166389] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] for vif in network_info: [ 828.166389] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 828.166389] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] return self._sync_wrapper(fn, *args, **kwargs) [ 828.166389] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 828.166389] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] self.wait() [ 828.166389] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 828.166389] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] self[:] = self._gt.wait() [ 828.166389] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 828.166389] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] return self._exit_event.wait() [ 828.166389] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 828.166389] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] current.throw(*self._exc) [ 828.166389] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 828.166389] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] result = function(*args, **kwargs) [ 828.166795] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 828.166795] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] return func(*args, **kwargs) [ 828.166795] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 828.166795] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] raise e [ 828.166795] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 828.166795] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] nwinfo = self.network_api.allocate_for_instance( [ 828.166795] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 828.166795] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] created_port_ids = self._update_ports_for_instance( [ 828.166795] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 828.166795] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] with excutils.save_and_reraise_exception(): [ 828.166795] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 828.166795] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] self.force_reraise() [ 828.166795] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 828.167315] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] raise self.value [ 828.167315] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 828.167315] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] updated_port = self._update_port( [ 828.167315] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 828.167315] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] _ensure_no_port_binding_failure(port) [ 828.167315] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 828.167315] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] raise exception.PortBindingFailed(port_id=port['id']) [ 828.167315] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] nova.exception.PortBindingFailed: Binding failed for port 04dda318-4cf1-44d4-bf27-fe6dce719865, please check neutron logs for more information. [ 828.167315] env[62753]: ERROR nova.compute.manager [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] [ 828.167315] env[62753]: DEBUG nova.compute.utils [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Binding failed for port 04dda318-4cf1-44d4-bf27-fe6dce719865, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 828.167645] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.910s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 828.168632] env[62753]: INFO nova.compute.claims [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 828.175017] env[62753]: DEBUG nova.compute.manager [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Build of instance 3adb208d-cf70-4d93-b034-a3e87d029f90 was re-scheduled: Binding failed for port 04dda318-4cf1-44d4-bf27-fe6dce719865, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 828.175017] env[62753]: DEBUG nova.compute.manager [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 828.175017] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Acquiring lock "refresh_cache-3adb208d-cf70-4d93-b034-a3e87d029f90" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.175017] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Acquired lock "refresh_cache-3adb208d-cf70-4d93-b034-a3e87d029f90" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.175354] env[62753]: DEBUG nova.network.neutron [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 828.340950] env[62753]: DEBUG nova.compute.manager [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 828.373628] env[62753]: DEBUG nova.virt.hardware [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 828.373967] env[62753]: DEBUG nova.virt.hardware [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 828.374045] env[62753]: DEBUG nova.virt.hardware [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 828.374202] env[62753]: DEBUG nova.virt.hardware [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 828.374348] env[62753]: DEBUG nova.virt.hardware [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 828.374493] env[62753]: DEBUG nova.virt.hardware [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 828.374748] env[62753]: DEBUG nova.virt.hardware [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 828.374889] env[62753]: DEBUG nova.virt.hardware [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 828.375069] env[62753]: DEBUG nova.virt.hardware [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 828.375305] env[62753]: DEBUG nova.virt.hardware [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 828.375419] env[62753]: DEBUG nova.virt.hardware [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 828.376367] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8eb33002-f3a5-4d5c-aa83-03f119d34bdb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.387282] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c471b99-0fef-4759-91ea-18caa0a679a6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 828.457106] env[62753]: ERROR nova.compute.manager [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 60dffaa7-7a7a-4d14-88e2-fc60b0015fd0, please check neutron logs for more information. [ 828.457106] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 828.457106] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 828.457106] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 828.457106] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 828.457106] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 828.457106] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 828.457106] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 828.457106] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 828.457106] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 828.457106] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 828.457106] env[62753]: ERROR nova.compute.manager raise self.value [ 828.457106] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 828.457106] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 828.457106] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 828.457106] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 828.457476] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 828.457476] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 828.457476] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 60dffaa7-7a7a-4d14-88e2-fc60b0015fd0, please check neutron logs for more information. [ 828.457476] env[62753]: ERROR nova.compute.manager [ 828.457476] env[62753]: Traceback (most recent call last): [ 828.457476] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 828.457476] env[62753]: listener.cb(fileno) [ 828.457476] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 828.457476] env[62753]: result = function(*args, **kwargs) [ 828.457476] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 828.457476] env[62753]: return func(*args, **kwargs) [ 828.457476] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 828.457476] env[62753]: raise e [ 828.457476] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 828.457476] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 828.457476] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 828.457476] env[62753]: created_port_ids = self._update_ports_for_instance( [ 828.457476] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 828.457476] env[62753]: with excutils.save_and_reraise_exception(): [ 828.457476] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 828.457476] env[62753]: self.force_reraise() [ 828.457476] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 828.457476] env[62753]: raise self.value [ 828.457476] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 828.457476] env[62753]: updated_port = self._update_port( [ 828.457476] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 828.457476] env[62753]: _ensure_no_port_binding_failure(port) [ 828.457476] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 828.457476] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 828.458073] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 60dffaa7-7a7a-4d14-88e2-fc60b0015fd0, please check neutron logs for more information. [ 828.458073] env[62753]: Removing descriptor: 18 [ 828.458073] env[62753]: ERROR nova.compute.manager [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 60dffaa7-7a7a-4d14-88e2-fc60b0015fd0, please check neutron logs for more information. [ 828.458073] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Traceback (most recent call last): [ 828.458073] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 828.458073] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] yield resources [ 828.458073] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 828.458073] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] self.driver.spawn(context, instance, image_meta, [ 828.458073] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 828.458073] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] self._vmops.spawn(context, instance, image_meta, injected_files, [ 828.458073] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 828.458073] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] vm_ref = self.build_virtual_machine(instance, [ 828.458406] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 828.458406] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] vif_infos = vmwarevif.get_vif_info(self._session, [ 828.458406] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 828.458406] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] for vif in network_info: [ 828.458406] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 828.458406] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] return self._sync_wrapper(fn, *args, **kwargs) [ 828.458406] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 828.458406] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] self.wait() [ 828.458406] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 828.458406] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] self[:] = self._gt.wait() [ 828.458406] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 828.458406] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] return self._exit_event.wait() [ 828.458406] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 828.458650] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] result = hub.switch() [ 828.458650] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 828.458650] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] return self.greenlet.switch() [ 828.458650] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 828.458650] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] result = function(*args, **kwargs) [ 828.458650] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 828.458650] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] return func(*args, **kwargs) [ 828.458650] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 828.458650] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] raise e [ 828.458650] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 828.458650] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] nwinfo = self.network_api.allocate_for_instance( [ 828.458650] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 828.458650] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] created_port_ids = self._update_ports_for_instance( [ 828.458890] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 828.458890] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] with excutils.save_and_reraise_exception(): [ 828.458890] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 828.458890] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] self.force_reraise() [ 828.458890] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 828.458890] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] raise self.value [ 828.458890] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 828.458890] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] updated_port = self._update_port( [ 828.458890] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 828.458890] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] _ensure_no_port_binding_failure(port) [ 828.458890] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 828.458890] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] raise exception.PortBindingFailed(port_id=port['id']) [ 828.459124] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] nova.exception.PortBindingFailed: Binding failed for port 60dffaa7-7a7a-4d14-88e2-fc60b0015fd0, please check neutron logs for more information. [ 828.459124] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] [ 828.459124] env[62753]: INFO nova.compute.manager [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Terminating instance [ 828.460925] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Acquiring lock "refresh_cache-7931ae4b-c486-4e57-9bf5-14166a1d4201" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 828.461149] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Acquired lock "refresh_cache-7931ae4b-c486-4e57-9bf5-14166a1d4201" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 828.461327] env[62753]: DEBUG nova.network.neutron [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 828.483142] env[62753]: DEBUG nova.network.neutron [-] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.695094] env[62753]: DEBUG nova.network.neutron [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.792046] env[62753]: DEBUG nova.network.neutron [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 828.980325] env[62753]: DEBUG nova.network.neutron [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 828.986666] env[62753]: INFO nova.compute.manager [-] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Took 1.06 seconds to deallocate network for instance. [ 828.989011] env[62753]: DEBUG nova.compute.claims [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 828.989259] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.068475] env[62753]: DEBUG nova.network.neutron [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.296208] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Releasing lock "refresh_cache-3adb208d-cf70-4d93-b034-a3e87d029f90" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.296435] env[62753]: DEBUG nova.compute.manager [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 829.296836] env[62753]: DEBUG nova.compute.manager [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 829.296836] env[62753]: DEBUG nova.network.neutron [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 829.314112] env[62753]: DEBUG nova.network.neutron [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 829.417785] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d89e8ab-1b33-4ef1-9543-a750e2fcc8ab {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.425792] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a28eabb9-ed47-49c1-b5fa-7322deccfac4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.454508] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ef10f29-2fb5-4da3-843e-1da0072d5048 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.461670] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1109c73-3cab-4c87-86ba-4952afc2d5cb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.474515] env[62753]: DEBUG nova.compute.provider_tree [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 829.575642] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Releasing lock "refresh_cache-7931ae4b-c486-4e57-9bf5-14166a1d4201" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 829.576145] env[62753]: DEBUG nova.compute.manager [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 829.576899] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 829.576899] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0009fdb3-ab8c-4d96-ac99-6d7732ad1b60 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.584944] env[62753]: DEBUG nova.compute.manager [req-e1089791-bc73-4853-8b8f-5e91818775b3 req-c379bf6b-dd72-4253-ab4e-bacee79da59e service nova] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Received event network-changed-60dffaa7-7a7a-4d14-88e2-fc60b0015fd0 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 829.585179] env[62753]: DEBUG nova.compute.manager [req-e1089791-bc73-4853-8b8f-5e91818775b3 req-c379bf6b-dd72-4253-ab4e-bacee79da59e service nova] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Refreshing instance network info cache due to event network-changed-60dffaa7-7a7a-4d14-88e2-fc60b0015fd0. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 829.586588] env[62753]: DEBUG oslo_concurrency.lockutils [req-e1089791-bc73-4853-8b8f-5e91818775b3 req-c379bf6b-dd72-4253-ab4e-bacee79da59e service nova] Acquiring lock "refresh_cache-7931ae4b-c486-4e57-9bf5-14166a1d4201" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 829.586588] env[62753]: DEBUG oslo_concurrency.lockutils [req-e1089791-bc73-4853-8b8f-5e91818775b3 req-c379bf6b-dd72-4253-ab4e-bacee79da59e service nova] Acquired lock "refresh_cache-7931ae4b-c486-4e57-9bf5-14166a1d4201" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 829.586588] env[62753]: DEBUG nova.network.neutron [req-e1089791-bc73-4853-8b8f-5e91818775b3 req-c379bf6b-dd72-4253-ab4e-bacee79da59e service nova] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Refreshing network info cache for port 60dffaa7-7a7a-4d14-88e2-fc60b0015fd0 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 829.591801] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f5ae911-f8c9-4a50-838f-59d1b7c6cea4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 829.613414] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 7931ae4b-c486-4e57-9bf5-14166a1d4201 could not be found. [ 829.613586] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 829.613758] env[62753]: INFO nova.compute.manager [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Took 0.04 seconds to destroy the instance on the hypervisor. [ 829.614057] env[62753]: DEBUG oslo.service.loopingcall [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 829.614277] env[62753]: DEBUG nova.compute.manager [-] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 829.614370] env[62753]: DEBUG nova.network.neutron [-] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 829.629951] env[62753]: DEBUG nova.network.neutron [-] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 829.749019] env[62753]: DEBUG oslo_concurrency.lockutils [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "50e1c1a1-7b8f-49cd-932d-03c920209634" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 829.749019] env[62753]: DEBUG oslo_concurrency.lockutils [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "50e1c1a1-7b8f-49cd-932d-03c920209634" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 829.817078] env[62753]: DEBUG nova.network.neutron [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 829.978276] env[62753]: DEBUG nova.scheduler.client.report [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 830.059664] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "33c1de85-aade-4c69-8fdb-7672d8c53300" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.059996] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "33c1de85-aade-4c69-8fdb-7672d8c53300" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.106013] env[62753]: DEBUG nova.network.neutron [req-e1089791-bc73-4853-8b8f-5e91818775b3 req-c379bf6b-dd72-4253-ab4e-bacee79da59e service nova] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 830.131894] env[62753]: DEBUG nova.network.neutron [-] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.190502] env[62753]: DEBUG nova.network.neutron [req-e1089791-bc73-4853-8b8f-5e91818775b3 req-c379bf6b-dd72-4253-ab4e-bacee79da59e service nova] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 830.319579] env[62753]: INFO nova.compute.manager [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] [instance: 3adb208d-cf70-4d93-b034-a3e87d029f90] Took 1.02 seconds to deallocate network for instance. [ 830.482383] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.315s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 830.482929] env[62753]: DEBUG nova.compute.manager [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 830.485618] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.666s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 830.636720] env[62753]: INFO nova.compute.manager [-] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Took 1.02 seconds to deallocate network for instance. [ 830.641061] env[62753]: DEBUG nova.compute.claims [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 830.641061] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 830.692955] env[62753]: DEBUG oslo_concurrency.lockutils [req-e1089791-bc73-4853-8b8f-5e91818775b3 req-c379bf6b-dd72-4253-ab4e-bacee79da59e service nova] Releasing lock "refresh_cache-7931ae4b-c486-4e57-9bf5-14166a1d4201" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 830.693205] env[62753]: DEBUG nova.compute.manager [req-e1089791-bc73-4853-8b8f-5e91818775b3 req-c379bf6b-dd72-4253-ab4e-bacee79da59e service nova] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Received event network-vif-deleted-60dffaa7-7a7a-4d14-88e2-fc60b0015fd0 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 830.993648] env[62753]: DEBUG nova.compute.utils [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 830.995819] env[62753]: DEBUG nova.compute.manager [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 830.995819] env[62753]: DEBUG nova.network.neutron [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 831.039884] env[62753]: DEBUG nova.policy [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '14bfb2a920864f73bfb5fc7a45e7bde8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '909db6176c8f475caf2562a593bc61a9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 831.234043] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ea63882-1358-4d70-9720-8fff6bc92823 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.240836] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33d23099-d458-4326-b174-26b30920be1c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.281321] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8da5d2e-d39e-4efe-9751-dbfd6d3060b2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.288721] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03084172-af09-4418-8a2a-5a760952c0ba {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 831.304085] env[62753]: DEBUG nova.compute.provider_tree [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 831.306019] env[62753]: DEBUG nova.network.neutron [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Successfully created port: 4a43717d-e663-4998-a93a-fc8c773710d3 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 831.346982] env[62753]: INFO nova.scheduler.client.report [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Deleted allocations for instance 3adb208d-cf70-4d93-b034-a3e87d029f90 [ 831.499467] env[62753]: DEBUG nova.compute.manager [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 831.809257] env[62753]: DEBUG nova.scheduler.client.report [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 831.854501] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3cf2419d-d151-4338-947a-62ef6efe9ae4 tempest-ImagesOneServerTestJSON-5320992 tempest-ImagesOneServerTestJSON-5320992-project-member] Lock "3adb208d-cf70-4d93-b034-a3e87d029f90" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.096s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.024788] env[62753]: DEBUG nova.compute.manager [req-ce5fd457-23d0-4777-838c-06fb7d74035c req-66f0d382-8d1a-4c00-a164-f2c7e62a446c service nova] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Received event network-changed-4a43717d-e663-4998-a93a-fc8c773710d3 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 832.025058] env[62753]: DEBUG nova.compute.manager [req-ce5fd457-23d0-4777-838c-06fb7d74035c req-66f0d382-8d1a-4c00-a164-f2c7e62a446c service nova] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Refreshing instance network info cache due to event network-changed-4a43717d-e663-4998-a93a-fc8c773710d3. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 832.025233] env[62753]: DEBUG oslo_concurrency.lockutils [req-ce5fd457-23d0-4777-838c-06fb7d74035c req-66f0d382-8d1a-4c00-a164-f2c7e62a446c service nova] Acquiring lock "refresh_cache-16c82d55-1420-4f83-a547-295e793de9df" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.025348] env[62753]: DEBUG oslo_concurrency.lockutils [req-ce5fd457-23d0-4777-838c-06fb7d74035c req-66f0d382-8d1a-4c00-a164-f2c7e62a446c service nova] Acquired lock "refresh_cache-16c82d55-1420-4f83-a547-295e793de9df" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.025506] env[62753]: DEBUG nova.network.neutron [req-ce5fd457-23d0-4777-838c-06fb7d74035c req-66f0d382-8d1a-4c00-a164-f2c7e62a446c service nova] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Refreshing network info cache for port 4a43717d-e663-4998-a93a-fc8c773710d3 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 832.186096] env[62753]: ERROR nova.compute.manager [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Instance failed network setup after 1 attempt(s): nova.exception.PortBindingFailed: Binding failed for port 4a43717d-e663-4998-a93a-fc8c773710d3, please check neutron logs for more information. [ 832.186096] env[62753]: ERROR nova.compute.manager Traceback (most recent call last): [ 832.186096] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 832.186096] env[62753]: ERROR nova.compute.manager nwinfo = self.network_api.allocate_for_instance( [ 832.186096] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 832.186096] env[62753]: ERROR nova.compute.manager created_port_ids = self._update_ports_for_instance( [ 832.186096] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 832.186096] env[62753]: ERROR nova.compute.manager with excutils.save_and_reraise_exception(): [ 832.186096] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 832.186096] env[62753]: ERROR nova.compute.manager self.force_reraise() [ 832.186096] env[62753]: ERROR nova.compute.manager File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 832.186096] env[62753]: ERROR nova.compute.manager raise self.value [ 832.186096] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 832.186096] env[62753]: ERROR nova.compute.manager updated_port = self._update_port( [ 832.186096] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 832.186096] env[62753]: ERROR nova.compute.manager _ensure_no_port_binding_failure(port) [ 832.186544] env[62753]: ERROR nova.compute.manager File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 832.186544] env[62753]: ERROR nova.compute.manager raise exception.PortBindingFailed(port_id=port['id']) [ 832.186544] env[62753]: ERROR nova.compute.manager nova.exception.PortBindingFailed: Binding failed for port 4a43717d-e663-4998-a93a-fc8c773710d3, please check neutron logs for more information. [ 832.186544] env[62753]: ERROR nova.compute.manager [ 832.186544] env[62753]: Traceback (most recent call last): [ 832.186544] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/poll.py", line 111, in wait [ 832.186544] env[62753]: listener.cb(fileno) [ 832.186544] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 832.186544] env[62753]: result = function(*args, **kwargs) [ 832.186544] env[62753]: File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 832.186544] env[62753]: return func(*args, **kwargs) [ 832.186544] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 832.186544] env[62753]: raise e [ 832.186544] env[62753]: File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 832.186544] env[62753]: nwinfo = self.network_api.allocate_for_instance( [ 832.186544] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 832.186544] env[62753]: created_port_ids = self._update_ports_for_instance( [ 832.186544] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 832.186544] env[62753]: with excutils.save_and_reraise_exception(): [ 832.186544] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 832.186544] env[62753]: self.force_reraise() [ 832.186544] env[62753]: File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 832.186544] env[62753]: raise self.value [ 832.186544] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 832.186544] env[62753]: updated_port = self._update_port( [ 832.186544] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 832.186544] env[62753]: _ensure_no_port_binding_failure(port) [ 832.186544] env[62753]: File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 832.186544] env[62753]: raise exception.PortBindingFailed(port_id=port['id']) [ 832.187476] env[62753]: nova.exception.PortBindingFailed: Binding failed for port 4a43717d-e663-4998-a93a-fc8c773710d3, please check neutron logs for more information. [ 832.187476] env[62753]: Removing descriptor: 18 [ 832.314263] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.828s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 832.314907] env[62753]: ERROR nova.compute.manager [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port f362843b-39bc-4173-8881-61ed1caca7c6, please check neutron logs for more information. [ 832.314907] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Traceback (most recent call last): [ 832.314907] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 832.314907] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] self.driver.spawn(context, instance, image_meta, [ 832.314907] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 832.314907] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] self._vmops.spawn(context, instance, image_meta, injected_files, [ 832.314907] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 832.314907] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] vm_ref = self.build_virtual_machine(instance, [ 832.314907] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 832.314907] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] vif_infos = vmwarevif.get_vif_info(self._session, [ 832.314907] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 832.315233] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] for vif in network_info: [ 832.315233] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 832.315233] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] return self._sync_wrapper(fn, *args, **kwargs) [ 832.315233] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 832.315233] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] self.wait() [ 832.315233] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 832.315233] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] self[:] = self._gt.wait() [ 832.315233] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 832.315233] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] return self._exit_event.wait() [ 832.315233] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 832.315233] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] current.throw(*self._exc) [ 832.315233] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 832.315233] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] result = function(*args, **kwargs) [ 832.315603] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 832.315603] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] return func(*args, **kwargs) [ 832.315603] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 832.315603] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] raise e [ 832.315603] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 832.315603] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] nwinfo = self.network_api.allocate_for_instance( [ 832.315603] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 832.315603] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] created_port_ids = self._update_ports_for_instance( [ 832.315603] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 832.315603] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] with excutils.save_and_reraise_exception(): [ 832.315603] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 832.315603] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] self.force_reraise() [ 832.315603] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 832.315958] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] raise self.value [ 832.315958] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 832.315958] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] updated_port = self._update_port( [ 832.315958] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 832.315958] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] _ensure_no_port_binding_failure(port) [ 832.315958] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 832.315958] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] raise exception.PortBindingFailed(port_id=port['id']) [ 832.315958] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] nova.exception.PortBindingFailed: Binding failed for port f362843b-39bc-4173-8881-61ed1caca7c6, please check neutron logs for more information. [ 832.315958] env[62753]: ERROR nova.compute.manager [instance: f63a8912-0abf-4c76-aaae-4085b3909688] [ 832.315958] env[62753]: DEBUG nova.compute.utils [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Binding failed for port f362843b-39bc-4173-8881-61ed1caca7c6, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 832.317024] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 23.128s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 832.322105] env[62753]: DEBUG nova.compute.manager [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Build of instance f63a8912-0abf-4c76-aaae-4085b3909688 was re-scheduled: Binding failed for port f362843b-39bc-4173-8881-61ed1caca7c6, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 832.322105] env[62753]: DEBUG nova.compute.manager [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 832.322105] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Acquiring lock "refresh_cache-f63a8912-0abf-4c76-aaae-4085b3909688" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.322105] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Acquired lock "refresh_cache-f63a8912-0abf-4c76-aaae-4085b3909688" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 832.322105] env[62753]: DEBUG nova.network.neutron [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 832.357495] env[62753]: DEBUG nova.compute.manager [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 832.508708] env[62753]: DEBUG nova.compute.manager [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 832.539939] env[62753]: DEBUG nova.virt.hardware [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 832.540193] env[62753]: DEBUG nova.virt.hardware [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 832.540408] env[62753]: DEBUG nova.virt.hardware [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 832.540731] env[62753]: DEBUG nova.virt.hardware [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 832.540907] env[62753]: DEBUG nova.virt.hardware [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 832.541070] env[62753]: DEBUG nova.virt.hardware [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 832.541285] env[62753]: DEBUG nova.virt.hardware [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 832.541443] env[62753]: DEBUG nova.virt.hardware [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 832.541607] env[62753]: DEBUG nova.virt.hardware [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 832.541766] env[62753]: DEBUG nova.virt.hardware [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 832.541935] env[62753]: DEBUG nova.virt.hardware [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 832.542828] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdea754a-1c1c-48e2-9ddb-ea8f881ca947 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.546187] env[62753]: DEBUG nova.network.neutron [req-ce5fd457-23d0-4777-838c-06fb7d74035c req-66f0d382-8d1a-4c00-a164-f2c7e62a446c service nova] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.553473] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701cf7aa-714f-4b32-915b-554fa149376d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 832.567548] env[62753]: ERROR nova.compute.manager [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Instance failed to spawn: nova.exception.PortBindingFailed: Binding failed for port 4a43717d-e663-4998-a93a-fc8c773710d3, please check neutron logs for more information. [ 832.567548] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] Traceback (most recent call last): [ 832.567548] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/compute/manager.py", line 2894, in _build_resources [ 832.567548] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] yield resources [ 832.567548] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 832.567548] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] self.driver.spawn(context, instance, image_meta, [ 832.567548] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 832.567548] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] self._vmops.spawn(context, instance, image_meta, injected_files, [ 832.567548] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 832.567548] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] vm_ref = self.build_virtual_machine(instance, [ 832.567548] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 832.567806] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] vif_infos = vmwarevif.get_vif_info(self._session, [ 832.567806] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 832.567806] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] for vif in network_info: [ 832.567806] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 832.567806] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] return self._sync_wrapper(fn, *args, **kwargs) [ 832.567806] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 832.567806] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] self.wait() [ 832.567806] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 832.567806] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] self[:] = self._gt.wait() [ 832.567806] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 832.567806] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] return self._exit_event.wait() [ 832.567806] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 832.567806] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] current.throw(*self._exc) [ 832.568123] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 832.568123] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] result = function(*args, **kwargs) [ 832.568123] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 832.568123] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] return func(*args, **kwargs) [ 832.568123] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 832.568123] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] raise e [ 832.568123] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 832.568123] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] nwinfo = self.network_api.allocate_for_instance( [ 832.568123] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 832.568123] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] created_port_ids = self._update_ports_for_instance( [ 832.568123] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 832.568123] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] with excutils.save_and_reraise_exception(): [ 832.568123] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 832.568389] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] self.force_reraise() [ 832.568389] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 832.568389] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] raise self.value [ 832.568389] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 832.568389] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] updated_port = self._update_port( [ 832.568389] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 832.568389] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] _ensure_no_port_binding_failure(port) [ 832.568389] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 832.568389] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] raise exception.PortBindingFailed(port_id=port['id']) [ 832.568389] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] nova.exception.PortBindingFailed: Binding failed for port 4a43717d-e663-4998-a93a-fc8c773710d3, please check neutron logs for more information. [ 832.568389] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] [ 832.568389] env[62753]: INFO nova.compute.manager [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Terminating instance [ 832.572019] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "refresh_cache-16c82d55-1420-4f83-a547-295e793de9df" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 832.620060] env[62753]: DEBUG nova.network.neutron [req-ce5fd457-23d0-4777-838c-06fb7d74035c req-66f0d382-8d1a-4c00-a164-f2c7e62a446c service nova] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.846858] env[62753]: DEBUG nova.network.neutron [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 832.913115] env[62753]: DEBUG nova.network.neutron [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 832.950401] env[62753]: DEBUG nova.scheduler.client.report [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Refreshing inventories for resource provider 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 832.973227] env[62753]: DEBUG nova.scheduler.client.report [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Updating ProviderTree inventory for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 832.974021] env[62753]: DEBUG nova.compute.provider_tree [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Updating inventory in ProviderTree for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 832.976292] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 832.984660] env[62753]: DEBUG nova.scheduler.client.report [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Refreshing aggregate associations for resource provider 42c8603e-76bc-4e29-ad77-cac0094f7308, aggregates: None {{(pid=62753) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 833.007696] env[62753]: DEBUG nova.scheduler.client.report [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Refreshing trait associations for resource provider 42c8603e-76bc-4e29-ad77-cac0094f7308, traits: COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62753) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 833.122777] env[62753]: DEBUG oslo_concurrency.lockutils [req-ce5fd457-23d0-4777-838c-06fb7d74035c req-66f0d382-8d1a-4c00-a164-f2c7e62a446c service nova] Releasing lock "refresh_cache-16c82d55-1420-4f83-a547-295e793de9df" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.123202] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquired lock "refresh_cache-16c82d55-1420-4f83-a547-295e793de9df" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 833.123389] env[62753]: DEBUG nova.network.neutron [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 833.238529] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0065109e-9038-4f3a-8a9a-e11734f9a390 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.246388] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46056375-4fcd-4144-b051-70112e71e292 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.278512] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6b8246f-385d-4af0-9453-cbfe10f0b132 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.285796] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a6807886-4b5b-4454-9249-3223978539db {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 833.298925] env[62753]: DEBUG nova.compute.provider_tree [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 833.415142] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Releasing lock "refresh_cache-f63a8912-0abf-4c76-aaae-4085b3909688" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 833.415485] env[62753]: DEBUG nova.compute.manager [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 833.415582] env[62753]: DEBUG nova.compute.manager [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 833.415718] env[62753]: DEBUG nova.network.neutron [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 833.433010] env[62753]: DEBUG nova.network.neutron [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.640538] env[62753]: DEBUG nova.network.neutron [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 833.718941] env[62753]: DEBUG nova.network.neutron [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 833.802078] env[62753]: DEBUG nova.scheduler.client.report [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 833.936106] env[62753]: DEBUG nova.network.neutron [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.048609] env[62753]: DEBUG nova.compute.manager [req-ee27c83f-8b13-4b27-af36-91878c9a0ba9 req-005f1810-db5c-4e17-be73-57b8231bb9f4 service nova] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Received event network-vif-deleted-4a43717d-e663-4998-a93a-fc8c773710d3 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 834.221479] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Releasing lock "refresh_cache-16c82d55-1420-4f83-a547-295e793de9df" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 834.222583] env[62753]: DEBUG nova.compute.manager [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 834.222583] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 834.222583] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4e4052bc-520f-4a2e-88fe-ff117687f3bd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.231747] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2377d820-0d98-4fd2-aad2-aa58beeafc1b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 834.254494] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 16c82d55-1420-4f83-a547-295e793de9df could not be found. [ 834.256035] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 834.256035] env[62753]: INFO nova.compute.manager [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Took 0.03 seconds to destroy the instance on the hypervisor. [ 834.256035] env[62753]: DEBUG oslo.service.loopingcall [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 834.256035] env[62753]: DEBUG nova.compute.manager [-] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 834.256035] env[62753]: DEBUG nova.network.neutron [-] [instance: 16c82d55-1420-4f83-a547-295e793de9df] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 834.273523] env[62753]: DEBUG nova.network.neutron [-] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.306920] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.990s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 834.307576] env[62753]: ERROR nova.compute.manager [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 33143ac6-b800-4c58-b336-0db7572c50a7, please check neutron logs for more information. [ 834.307576] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Traceback (most recent call last): [ 834.307576] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 834.307576] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] self.driver.spawn(context, instance, image_meta, [ 834.307576] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 834.307576] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] self._vmops.spawn(context, instance, image_meta, injected_files, [ 834.307576] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 834.307576] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] vm_ref = self.build_virtual_machine(instance, [ 834.307576] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 834.307576] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] vif_infos = vmwarevif.get_vif_info(self._session, [ 834.307576] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 834.307892] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] for vif in network_info: [ 834.307892] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 834.307892] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] return self._sync_wrapper(fn, *args, **kwargs) [ 834.307892] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 834.307892] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] self.wait() [ 834.307892] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 834.307892] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] self[:] = self._gt.wait() [ 834.307892] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 834.307892] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] return self._exit_event.wait() [ 834.307892] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 834.307892] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] current.throw(*self._exc) [ 834.307892] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 834.307892] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] result = function(*args, **kwargs) [ 834.308232] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 834.308232] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] return func(*args, **kwargs) [ 834.308232] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 834.308232] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] raise e [ 834.308232] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 834.308232] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] nwinfo = self.network_api.allocate_for_instance( [ 834.308232] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 834.308232] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] created_port_ids = self._update_ports_for_instance( [ 834.308232] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 834.308232] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] with excutils.save_and_reraise_exception(): [ 834.308232] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 834.308232] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] self.force_reraise() [ 834.308232] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 834.308570] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] raise self.value [ 834.308570] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 834.308570] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] updated_port = self._update_port( [ 834.308570] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 834.308570] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] _ensure_no_port_binding_failure(port) [ 834.308570] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 834.308570] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] raise exception.PortBindingFailed(port_id=port['id']) [ 834.308570] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] nova.exception.PortBindingFailed: Binding failed for port 33143ac6-b800-4c58-b336-0db7572c50a7, please check neutron logs for more information. [ 834.308570] env[62753]: ERROR nova.compute.manager [instance: 7563bf4e-39c5-4831-a415-cb311e237762] [ 834.308570] env[62753]: DEBUG nova.compute.utils [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Binding failed for port 33143ac6-b800-4c58-b336-0db7572c50a7, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 834.309630] env[62753]: DEBUG oslo_concurrency.lockutils [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 20.906s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 834.309807] env[62753]: DEBUG nova.objects.instance [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62753) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 834.313194] env[62753]: DEBUG nova.compute.manager [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Build of instance 7563bf4e-39c5-4831-a415-cb311e237762 was re-scheduled: Binding failed for port 33143ac6-b800-4c58-b336-0db7572c50a7, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 834.313603] env[62753]: DEBUG nova.compute.manager [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 834.314021] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "refresh_cache-7563bf4e-39c5-4831-a415-cb311e237762" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 834.314121] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "refresh_cache-7563bf4e-39c5-4831-a415-cb311e237762" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 834.314280] env[62753]: DEBUG nova.network.neutron [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 834.440222] env[62753]: INFO nova.compute.manager [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] [instance: f63a8912-0abf-4c76-aaae-4085b3909688] Took 1.02 seconds to deallocate network for instance. [ 834.775886] env[62753]: DEBUG nova.network.neutron [-] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 834.836447] env[62753]: DEBUG nova.network.neutron [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 834.916835] env[62753]: DEBUG nova.network.neutron [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.278566] env[62753]: INFO nova.compute.manager [-] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Took 1.02 seconds to deallocate network for instance. [ 835.280842] env[62753]: DEBUG nova.compute.claims [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Aborting claim: {{(pid=62753) abort /opt/stack/nova/nova/compute/claims.py:85}} [ 835.281036] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 835.323427] env[62753]: DEBUG oslo_concurrency.lockutils [None req-29bd314c-0bc9-40f1-887d-adde2e1508ed tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.014s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 835.324740] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 13.865s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 835.421209] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "refresh_cache-7563bf4e-39c5-4831-a415-cb311e237762" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 835.421451] env[62753]: DEBUG nova.compute.manager [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 835.421683] env[62753]: DEBUG nova.compute.manager [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 835.421957] env[62753]: DEBUG nova.network.neutron [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 835.441531] env[62753]: DEBUG nova.network.neutron [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 835.477484] env[62753]: INFO nova.scheduler.client.report [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Deleted allocations for instance f63a8912-0abf-4c76-aaae-4085b3909688 [ 835.945564] env[62753]: DEBUG nova.network.neutron [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 835.988273] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6427c7cb-448a-48f9-a17b-0b1c63261916 tempest-MigrationsAdminTest-6628426 tempest-MigrationsAdminTest-6628426-project-member] Lock "f63a8912-0abf-4c76-aaae-4085b3909688" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 146.504s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 836.075138] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0ad642e-9f5c-429d-a9e4-9b738240658d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.086645] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f1d48bc-8f9a-4944-abe2-46332587d63c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.115479] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e7ae36-4dd5-4951-a81e-da04f7972a61 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.123336] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46de3b95-bc41-45cb-ac1f-83b83d8f2cc5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 836.136565] env[62753]: DEBUG nova.compute.provider_tree [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 836.448030] env[62753]: INFO nova.compute.manager [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 7563bf4e-39c5-4831-a415-cb311e237762] Took 1.03 seconds to deallocate network for instance. [ 836.493697] env[62753]: DEBUG nova.compute.manager [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 836.639789] env[62753]: DEBUG nova.scheduler.client.report [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 837.019033] env[62753]: DEBUG oslo_concurrency.lockutils [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 837.146090] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.821s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 837.146694] env[62753]: ERROR nova.compute.manager [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 61585ec2-05bb-498f-b46f-b26cd54d0f0c, please check neutron logs for more information. [ 837.146694] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Traceback (most recent call last): [ 837.146694] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 837.146694] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] self.driver.spawn(context, instance, image_meta, [ 837.146694] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 837.146694] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] self._vmops.spawn(context, instance, image_meta, injected_files, [ 837.146694] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 837.146694] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] vm_ref = self.build_virtual_machine(instance, [ 837.146694] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 837.146694] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] vif_infos = vmwarevif.get_vif_info(self._session, [ 837.146694] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 837.146971] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] for vif in network_info: [ 837.146971] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 837.146971] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] return self._sync_wrapper(fn, *args, **kwargs) [ 837.146971] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 837.146971] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] self.wait() [ 837.146971] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 837.146971] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] self[:] = self._gt.wait() [ 837.146971] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 837.146971] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] return self._exit_event.wait() [ 837.146971] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 837.146971] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] current.throw(*self._exc) [ 837.146971] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 837.146971] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] result = function(*args, **kwargs) [ 837.147257] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 837.147257] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] return func(*args, **kwargs) [ 837.147257] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 837.147257] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] raise e [ 837.147257] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 837.147257] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] nwinfo = self.network_api.allocate_for_instance( [ 837.147257] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 837.147257] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] created_port_ids = self._update_ports_for_instance( [ 837.147257] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 837.147257] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] with excutils.save_and_reraise_exception(): [ 837.147257] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 837.147257] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] self.force_reraise() [ 837.147257] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 837.147546] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] raise self.value [ 837.147546] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 837.147546] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] updated_port = self._update_port( [ 837.147546] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 837.147546] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] _ensure_no_port_binding_failure(port) [ 837.147546] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 837.147546] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] raise exception.PortBindingFailed(port_id=port['id']) [ 837.147546] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] nova.exception.PortBindingFailed: Binding failed for port 61585ec2-05bb-498f-b46f-b26cd54d0f0c, please check neutron logs for more information. [ 837.147546] env[62753]: ERROR nova.compute.manager [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] [ 837.147546] env[62753]: DEBUG nova.compute.utils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Binding failed for port 61585ec2-05bb-498f-b46f-b26cd54d0f0c, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 837.148621] env[62753]: DEBUG oslo_concurrency.lockutils [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.066s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 837.150693] env[62753]: INFO nova.compute.claims [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 837.153540] env[62753]: DEBUG nova.compute.manager [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Build of instance 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07 was re-scheduled: Binding failed for port 61585ec2-05bb-498f-b46f-b26cd54d0f0c, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 837.154094] env[62753]: DEBUG nova.compute.manager [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 837.154253] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Acquiring lock "refresh_cache-07c4e02b-a8ca-41bb-82ec-38a81b9f1d07" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 837.154400] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Acquired lock "refresh_cache-07c4e02b-a8ca-41bb-82ec-38a81b9f1d07" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 837.154567] env[62753]: DEBUG nova.network.neutron [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 837.482035] env[62753]: INFO nova.scheduler.client.report [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Deleted allocations for instance 7563bf4e-39c5-4831-a415-cb311e237762 [ 837.678076] env[62753]: DEBUG nova.network.neutron [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 837.770072] env[62753]: DEBUG nova.network.neutron [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 837.997687] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9334e787-4caa-4fe6-af53-6e2e3a6f32e6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "7563bf4e-39c5-4831-a415-cb311e237762" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 147.202s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 838.275573] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Releasing lock "refresh_cache-07c4e02b-a8ca-41bb-82ec-38a81b9f1d07" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 838.275808] env[62753]: DEBUG nova.compute.manager [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 838.275985] env[62753]: DEBUG nova.compute.manager [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 838.276162] env[62753]: DEBUG nova.network.neutron [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 838.296500] env[62753]: DEBUG nova.network.neutron [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 838.437660] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b5f60b-b27c-4560-a52d-01b13349a430 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.446685] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666cbc5b-f933-46a9-ab49-c614508dc375 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.480523] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-287e3a05-d932-4963-982e-01119fe7b217 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.486212] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Acquiring lock "57eef548-a0f1-4f0b-a026-885a10d005c2" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 838.486488] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Lock "57eef548-a0f1-4f0b-a026-885a10d005c2" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 838.491922] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d5addd8-cf28-4a67-ba80-4009524234f9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 838.506728] env[62753]: DEBUG nova.compute.manager [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 838.510178] env[62753]: DEBUG nova.compute.provider_tree [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 838.799439] env[62753]: DEBUG nova.network.neutron [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 839.604200] env[62753]: DEBUG nova.scheduler.client.report [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 839.607553] env[62753]: INFO nova.compute.manager [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] [instance: 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07] Took 1.33 seconds to deallocate network for instance. [ 839.615162] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "f03739f4-921f-46e5-b0e7-e9a99d96d621" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 839.615162] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "f03739f4-921f-46e5-b0e7-e9a99d96d621" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 839.628366] env[62753]: DEBUG oslo_concurrency.lockutils [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 840.116829] env[62753]: DEBUG oslo_concurrency.lockutils [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.968s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 840.117602] env[62753]: DEBUG nova.compute.manager [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 840.127942] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 17.638s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 840.127942] env[62753]: DEBUG nova.objects.instance [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62753) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 840.637454] env[62753]: DEBUG nova.compute.utils [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 840.638686] env[62753]: DEBUG nova.compute.manager [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 840.638865] env[62753]: DEBUG nova.network.neutron [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 840.646032] env[62753]: INFO nova.scheduler.client.report [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Deleted allocations for instance 07c4e02b-a8ca-41bb-82ec-38a81b9f1d07 [ 840.693440] env[62753]: DEBUG nova.policy [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9931b40cfd7846038805c6a4caedaac0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b8e81660b30408c998e412f5fa81469', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 840.985576] env[62753]: DEBUG nova.network.neutron [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Successfully created port: 7e8e502a-7a64-4dfd-8f9d-18965ac19b6b {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 841.140408] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a2c68d-d678-4c98-aa2f-b6508631ff1a tempest-ServersAdmin275Test-2002162078 tempest-ServersAdmin275Test-2002162078-project-admin] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.018s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.144900] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 16.982s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 841.148112] env[62753]: DEBUG nova.compute.manager [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 841.155162] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f45488cc-124d-4fda-87a9-3796d21d1b11 tempest-ServerGroupTestJSON-241167154 tempest-ServerGroupTestJSON-241167154-project-member] Lock "07c4e02b-a8ca-41bb-82ec-38a81b9f1d07" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 148.267s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 841.658591] env[62753]: DEBUG nova.compute.manager [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 841.869821] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06a5c37e-fa95-4353-b498-0d74709faf81 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.881024] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15ff93df-6618-46d7-a95c-471f24ab210c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.907993] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c2282441-c0f0-481a-bce8-d04463e54648 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.917901] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d25f6c73-371c-4d5a-a75a-2603ea6ee41f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 841.934342] env[62753]: DEBUG nova.compute.provider_tree [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 842.162640] env[62753]: DEBUG nova.compute.manager [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 842.188347] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.192552] env[62753]: DEBUG nova.virt.hardware [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 842.192552] env[62753]: DEBUG nova.virt.hardware [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 842.192552] env[62753]: DEBUG nova.virt.hardware [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 842.192731] env[62753]: DEBUG nova.virt.hardware [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 842.192731] env[62753]: DEBUG nova.virt.hardware [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 842.192731] env[62753]: DEBUG nova.virt.hardware [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 842.192807] env[62753]: DEBUG nova.virt.hardware [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 842.192914] env[62753]: DEBUG nova.virt.hardware [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 842.193087] env[62753]: DEBUG nova.virt.hardware [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 842.193477] env[62753]: DEBUG nova.virt.hardware [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 842.193477] env[62753]: DEBUG nova.virt.hardware [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 842.194355] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-981bd0be-072e-479f-97f4-346f71a0b53f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.202770] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17df7d95-8e95-4b4d-85aa-baf7382ce080 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.437079] env[62753]: DEBUG nova.scheduler.client.report [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 842.686068] env[62753]: DEBUG nova.compute.manager [req-ca2351a9-b6be-41d9-acca-34e47b9f9fee req-45e2d1ca-961b-457a-b400-929df7e9ef99 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Received event network-vif-plugged-7e8e502a-7a64-4dfd-8f9d-18965ac19b6b {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 842.686318] env[62753]: DEBUG oslo_concurrency.lockutils [req-ca2351a9-b6be-41d9-acca-34e47b9f9fee req-45e2d1ca-961b-457a-b400-929df7e9ef99 service nova] Acquiring lock "3bc05fdb-81be-4764-b9d9-01acd125b020-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 842.686499] env[62753]: DEBUG oslo_concurrency.lockutils [req-ca2351a9-b6be-41d9-acca-34e47b9f9fee req-45e2d1ca-961b-457a-b400-929df7e9ef99 service nova] Lock "3bc05fdb-81be-4764-b9d9-01acd125b020-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.686745] env[62753]: DEBUG oslo_concurrency.lockutils [req-ca2351a9-b6be-41d9-acca-34e47b9f9fee req-45e2d1ca-961b-457a-b400-929df7e9ef99 service nova] Lock "3bc05fdb-81be-4764-b9d9-01acd125b020-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.686929] env[62753]: DEBUG nova.compute.manager [req-ca2351a9-b6be-41d9-acca-34e47b9f9fee req-45e2d1ca-961b-457a-b400-929df7e9ef99 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] No waiting events found dispatching network-vif-plugged-7e8e502a-7a64-4dfd-8f9d-18965ac19b6b {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 842.688782] env[62753]: WARNING nova.compute.manager [req-ca2351a9-b6be-41d9-acca-34e47b9f9fee req-45e2d1ca-961b-457a-b400-929df7e9ef99 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Received unexpected event network-vif-plugged-7e8e502a-7a64-4dfd-8f9d-18965ac19b6b for instance with vm_state building and task_state spawning. [ 842.857216] env[62753]: DEBUG nova.network.neutron [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Successfully updated port: 7e8e502a-7a64-4dfd-8f9d-18965ac19b6b {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 842.942459] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.801s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.943133] env[62753]: ERROR nova.compute.manager [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port a0fe6d5a-8a68-4ab9-8d30-b29003f609be, please check neutron logs for more information. [ 842.943133] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Traceback (most recent call last): [ 842.943133] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 842.943133] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] self.driver.spawn(context, instance, image_meta, [ 842.943133] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 842.943133] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] self._vmops.spawn(context, instance, image_meta, injected_files, [ 842.943133] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 842.943133] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] vm_ref = self.build_virtual_machine(instance, [ 842.943133] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 842.943133] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] vif_infos = vmwarevif.get_vif_info(self._session, [ 842.943133] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 842.943417] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] for vif in network_info: [ 842.943417] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 842.943417] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] return self._sync_wrapper(fn, *args, **kwargs) [ 842.943417] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 842.943417] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] self.wait() [ 842.943417] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 842.943417] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] self[:] = self._gt.wait() [ 842.943417] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 842.943417] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] return self._exit_event.wait() [ 842.943417] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 842.943417] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] current.throw(*self._exc) [ 842.943417] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 842.943417] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] result = function(*args, **kwargs) [ 842.943725] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 842.943725] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] return func(*args, **kwargs) [ 842.943725] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 842.943725] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] raise e [ 842.943725] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 842.943725] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] nwinfo = self.network_api.allocate_for_instance( [ 842.943725] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 842.943725] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] created_port_ids = self._update_ports_for_instance( [ 842.943725] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 842.943725] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] with excutils.save_and_reraise_exception(): [ 842.943725] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 842.943725] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] self.force_reraise() [ 842.943725] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 842.944044] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] raise self.value [ 842.944044] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 842.944044] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] updated_port = self._update_port( [ 842.944044] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 842.944044] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] _ensure_no_port_binding_failure(port) [ 842.944044] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 842.944044] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] raise exception.PortBindingFailed(port_id=port['id']) [ 842.944044] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] nova.exception.PortBindingFailed: Binding failed for port a0fe6d5a-8a68-4ab9-8d30-b29003f609be, please check neutron logs for more information. [ 842.944044] env[62753]: ERROR nova.compute.manager [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] [ 842.944044] env[62753]: DEBUG nova.compute.utils [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Binding failed for port a0fe6d5a-8a68-4ab9-8d30-b29003f609be, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 842.945129] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 17.409s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.945312] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 842.945465] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62753) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 842.945745] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.559s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 842.947187] env[62753]: INFO nova.compute.claims [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 842.949825] env[62753]: DEBUG nova.compute.manager [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Build of instance 67f5d96c-b870-442a-a0f3-a70533c045ea was re-scheduled: Binding failed for port a0fe6d5a-8a68-4ab9-8d30-b29003f609be, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 842.950277] env[62753]: DEBUG nova.compute.manager [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 842.950523] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "refresh_cache-67f5d96c-b870-442a-a0f3-a70533c045ea" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 842.950637] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "refresh_cache-67f5d96c-b870-442a-a0f3-a70533c045ea" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 842.950794] env[62753]: DEBUG nova.network.neutron [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 842.952275] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d51eb5-0e27-4558-8dd5-349e1212872b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.962921] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09a08c54-87a5-40b9-b5b5-9bad044ce0e8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.977360] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-930eeccd-cb39-4583-aa9b-9bbc31206fb8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 842.985157] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19a0aab1-e645-4e8a-8ad0-6a8bf88b034a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 843.013360] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181415MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=62753) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 843.013486] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 843.359951] env[62753]: DEBUG oslo_concurrency.lockutils [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 843.360126] env[62753]: DEBUG oslo_concurrency.lockutils [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 843.360293] env[62753]: DEBUG nova.network.neutron [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 843.478236] env[62753]: DEBUG nova.network.neutron [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 843.623458] env[62753]: DEBUG nova.network.neutron [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 843.926772] env[62753]: DEBUG nova.network.neutron [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.129018] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "refresh_cache-67f5d96c-b870-442a-a0f3-a70533c045ea" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.129018] env[62753]: DEBUG nova.compute.manager [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 844.129018] env[62753]: DEBUG nova.compute.manager [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 844.129018] env[62753]: DEBUG nova.network.neutron [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 844.158554] env[62753]: DEBUG nova.network.neutron [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 844.213737] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55f310cc-62a8-4482-80be-6a61444c1b4f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.224241] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1fdda6a-06f6-463b-b85b-75e94749fc56 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.256734] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bcb896d-2c51-49b6-92e5-b9215afefbb4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.264817] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622e142a-b28d-4106-a7bb-b6fc11e2848d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.280733] env[62753]: DEBUG nova.compute.provider_tree [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 844.315686] env[62753]: DEBUG nova.network.neutron [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Updating instance_info_cache with network_info: [{"id": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "address": "fa:16:3e:1d:84:1b", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8e502a-7a", "ovs_interfaceid": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.667784] env[62753]: DEBUG nova.network.neutron [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 844.725055] env[62753]: DEBUG nova.compute.manager [req-f24effb8-a38b-4ab3-98d2-b907915fe915 req-902bdb9d-fbd3-4f80-a683-c8ad7c63b92c service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Received event network-changed-7e8e502a-7a64-4dfd-8f9d-18965ac19b6b {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 844.725055] env[62753]: DEBUG nova.compute.manager [req-f24effb8-a38b-4ab3-98d2-b907915fe915 req-902bdb9d-fbd3-4f80-a683-c8ad7c63b92c service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Refreshing instance network info cache due to event network-changed-7e8e502a-7a64-4dfd-8f9d-18965ac19b6b. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 844.725055] env[62753]: DEBUG oslo_concurrency.lockutils [req-f24effb8-a38b-4ab3-98d2-b907915fe915 req-902bdb9d-fbd3-4f80-a683-c8ad7c63b92c service nova] Acquiring lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 844.784896] env[62753]: DEBUG nova.scheduler.client.report [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 844.819677] env[62753]: DEBUG oslo_concurrency.lockutils [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 844.820062] env[62753]: DEBUG nova.compute.manager [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Instance network_info: |[{"id": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "address": "fa:16:3e:1d:84:1b", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8e502a-7a", "ovs_interfaceid": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 844.820411] env[62753]: DEBUG oslo_concurrency.lockutils [req-f24effb8-a38b-4ab3-98d2-b907915fe915 req-902bdb9d-fbd3-4f80-a683-c8ad7c63b92c service nova] Acquired lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 844.820638] env[62753]: DEBUG nova.network.neutron [req-f24effb8-a38b-4ab3-98d2-b907915fe915 req-902bdb9d-fbd3-4f80-a683-c8ad7c63b92c service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Refreshing network info cache for port 7e8e502a-7a64-4dfd-8f9d-18965ac19b6b {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 844.821905] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:1d:84:1b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd48f0ef6-34e5-44d4-8baf-4470ed96ce73', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7e8e502a-7a64-4dfd-8f9d-18965ac19b6b', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 844.830729] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Creating folder: Project (7b8e81660b30408c998e412f5fa81469). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.831967] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-167e4f35-e1ec-4641-9cc2-6061bc8ed574 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.843255] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Created folder: Project (7b8e81660b30408c998e412f5fa81469) in parent group-v284541. [ 844.843445] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Creating folder: Instances. Parent ref: group-v284561. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 844.843683] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a4094fc6-b671-4949-8879-c78aa0a30624 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.854295] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Created folder: Instances in parent group-v284561. [ 844.854295] env[62753]: DEBUG oslo.service.loopingcall [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 844.854450] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 844.854536] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-49cafff6-bfd2-4ccf-8fa0-ce5375a3cb48 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 844.874165] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 844.874165] env[62753]: value = "task-1332164" [ 844.874165] env[62753]: _type = "Task" [ 844.874165] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 844.881485] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332164, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.174097] env[62753]: INFO nova.compute.manager [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 67f5d96c-b870-442a-a0f3-a70533c045ea] Took 1.05 seconds to deallocate network for instance. [ 845.290634] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 845.291373] env[62753]: DEBUG nova.compute.manager [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 845.296628] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.948s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 845.296628] env[62753]: DEBUG nova.objects.instance [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Lazy-loading 'resources' on Instance uuid 41476adb-95e5-4617-b178-9c981d13c43d {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 845.386732] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332164, 'name': CreateVM_Task, 'duration_secs': 0.391607} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.386911] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 845.396768] env[62753]: DEBUG oslo_concurrency.lockutils [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.396768] env[62753]: DEBUG oslo_concurrency.lockutils [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.396768] env[62753]: DEBUG oslo_concurrency.lockutils [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 845.397372] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b06aec8-7614-46c5-a2bd-95107988a898 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.402488] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 845.402488] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dfe820-b673-d799-b986-ed1e3c514b5d" [ 845.402488] env[62753]: _type = "Task" [ 845.402488] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.411629] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dfe820-b673-d799-b986-ed1e3c514b5d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 845.662299] env[62753]: DEBUG nova.network.neutron [req-f24effb8-a38b-4ab3-98d2-b907915fe915 req-902bdb9d-fbd3-4f80-a683-c8ad7c63b92c service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Updated VIF entry in instance network info cache for port 7e8e502a-7a64-4dfd-8f9d-18965ac19b6b. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 845.662707] env[62753]: DEBUG nova.network.neutron [req-f24effb8-a38b-4ab3-98d2-b907915fe915 req-902bdb9d-fbd3-4f80-a683-c8ad7c63b92c service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Updating instance_info_cache with network_info: [{"id": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "address": "fa:16:3e:1d:84:1b", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8e502a-7a", "ovs_interfaceid": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 845.799268] env[62753]: DEBUG nova.compute.utils [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 845.800635] env[62753]: DEBUG nova.compute.manager [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 845.800970] env[62753]: DEBUG nova.network.neutron [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 845.862855] env[62753]: DEBUG nova.policy [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cbc9dc744e144ff28cf4c44fb043094f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e376836411ae4854965adf2923eaaffd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 845.918058] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dfe820-b673-d799-b986-ed1e3c514b5d, 'name': SearchDatastore_Task, 'duration_secs': 0.014191} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 845.918058] env[62753]: DEBUG oslo_concurrency.lockutils [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 845.918058] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 845.918171] env[62753]: DEBUG oslo_concurrency.lockutils [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 845.918763] env[62753]: DEBUG oslo_concurrency.lockutils [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 845.918763] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 845.918763] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f2de41c2-26d0-4d82-aa58-d621c4003449 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.931331] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 845.931331] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 845.931331] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b6d15dc-d00c-4d5a-941d-f2ceea6d967a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 845.940399] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 845.940399] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ca5c00-9994-4be6-d999-60c90d9d692f" [ 845.940399] env[62753]: _type = "Task" [ 845.940399] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 845.948947] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ca5c00-9994-4be6-d999-60c90d9d692f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.119503] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-302513f0-597c-4054-9678-3a9a3da60324 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.129486] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ad49868-6746-4555-a4d5-d514e75c1e85 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.164262] env[62753]: DEBUG nova.network.neutron [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Successfully created port: ef0eb4ad-a83b-435a-a813-4434a8616c14 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 846.166614] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-277ba0ef-2748-463c-ae42-eab8036e2a05 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.170058] env[62753]: DEBUG oslo_concurrency.lockutils [req-f24effb8-a38b-4ab3-98d2-b907915fe915 req-902bdb9d-fbd3-4f80-a683-c8ad7c63b92c service nova] Releasing lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.177061] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0af98874-cd08-4297-bf41-3eb7b361b04c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.196706] env[62753]: DEBUG nova.compute.provider_tree [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 846.228887] env[62753]: INFO nova.scheduler.client.report [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Deleted allocations for instance 67f5d96c-b870-442a-a0f3-a70533c045ea [ 846.308252] env[62753]: DEBUG nova.compute.manager [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 846.453863] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ca5c00-9994-4be6-d999-60c90d9d692f, 'name': SearchDatastore_Task, 'duration_secs': 0.013415} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.454370] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af62d1e0-9665-4eda-9f86-941a34ede5e8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.459720] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 846.459720] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e0d926-162b-08b7-f726-aedc5cbdb638" [ 846.459720] env[62753]: _type = "Task" [ 846.459720] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.468186] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e0d926-162b-08b7-f726-aedc5cbdb638, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 846.700609] env[62753]: DEBUG nova.scheduler.client.report [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 846.738097] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d50d8ee9-d62c-4164-8949-47ecca117d1e tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "67f5d96c-b870-442a-a0f3-a70533c045ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 152.541s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 846.889035] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Acquiring lock "9a8a148c-057d-4c91-8b7c-4e05e6ac0f66" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 846.889167] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Lock "9a8a148c-057d-4c91-8b7c-4e05e6ac0f66" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 846.970896] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e0d926-162b-08b7-f726-aedc5cbdb638, 'name': SearchDatastore_Task, 'duration_secs': 0.009338} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 846.971151] env[62753]: DEBUG oslo_concurrency.lockutils [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 846.971396] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 3bc05fdb-81be-4764-b9d9-01acd125b020/3bc05fdb-81be-4764-b9d9-01acd125b020.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 846.971634] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-18f93da3-fa15-4eef-8dcc-fd5ac4eba34c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 846.977847] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 846.977847] env[62753]: value = "task-1332165" [ 846.977847] env[62753]: _type = "Task" [ 846.977847] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 846.985799] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332165, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.207389] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.912s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.208865] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.219s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.238696] env[62753]: INFO nova.scheduler.client.report [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Deleted allocations for instance 41476adb-95e5-4617-b178-9c981d13c43d [ 847.239976] env[62753]: DEBUG nova.compute.manager [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 847.320010] env[62753]: DEBUG nova.compute.manager [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 847.344651] env[62753]: DEBUG nova.virt.hardware [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 847.345185] env[62753]: DEBUG nova.virt.hardware [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 847.345391] env[62753]: DEBUG nova.virt.hardware [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 847.345639] env[62753]: DEBUG nova.virt.hardware [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 847.345737] env[62753]: DEBUG nova.virt.hardware [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 847.345870] env[62753]: DEBUG nova.virt.hardware [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 847.346105] env[62753]: DEBUG nova.virt.hardware [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 847.346271] env[62753]: DEBUG nova.virt.hardware [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 847.346439] env[62753]: DEBUG nova.virt.hardware [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 847.346601] env[62753]: DEBUG nova.virt.hardware [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 847.346815] env[62753]: DEBUG nova.virt.hardware [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 847.347673] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51356a58-209f-4b4f-ae07-f8d130a5b47a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.356182] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c8f407c-f3d4-44e9-9c1e-0c253c74f176 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.489935] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332165, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.481956} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 847.490431] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 3bc05fdb-81be-4764-b9d9-01acd125b020/3bc05fdb-81be-4764-b9d9-01acd125b020.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 847.490737] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 847.491214] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-61833a1d-707a-43ef-9c28-0434512e703a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 847.500428] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 847.500428] env[62753]: value = "task-1332166" [ 847.500428] env[62753]: _type = "Task" [ 847.500428] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 847.511975] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332166, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 847.758821] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6ae75a6d-0980-41b2-b470-a238306c10a8 tempest-ServersAdmin275Test-193143639 tempest-ServersAdmin275Test-193143639-project-member] Lock "41476adb-95e5-4617-b178-9c981d13c43d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 24.114s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.779175] env[62753]: DEBUG oslo_concurrency.lockutils [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.836848] env[62753]: DEBUG nova.compute.manager [req-0c1e76ea-32da-4eaf-a72a-a1287bc9c909 req-35b66b9f-82e6-4eea-b875-f620af015da0 service nova] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Received event network-vif-plugged-ef0eb4ad-a83b-435a-a813-4434a8616c14 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 847.836973] env[62753]: DEBUG oslo_concurrency.lockutils [req-0c1e76ea-32da-4eaf-a72a-a1287bc9c909 req-35b66b9f-82e6-4eea-b875-f620af015da0 service nova] Acquiring lock "046fc3dc-a187-4fda-bc66-345e1226f83d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 847.837187] env[62753]: DEBUG oslo_concurrency.lockutils [req-0c1e76ea-32da-4eaf-a72a-a1287bc9c909 req-35b66b9f-82e6-4eea-b875-f620af015da0 service nova] Lock "046fc3dc-a187-4fda-bc66-345e1226f83d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 847.837320] env[62753]: DEBUG oslo_concurrency.lockutils [req-0c1e76ea-32da-4eaf-a72a-a1287bc9c909 req-35b66b9f-82e6-4eea-b875-f620af015da0 service nova] Lock "046fc3dc-a187-4fda-bc66-345e1226f83d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 847.837472] env[62753]: DEBUG nova.compute.manager [req-0c1e76ea-32da-4eaf-a72a-a1287bc9c909 req-35b66b9f-82e6-4eea-b875-f620af015da0 service nova] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] No waiting events found dispatching network-vif-plugged-ef0eb4ad-a83b-435a-a813-4434a8616c14 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 847.837617] env[62753]: WARNING nova.compute.manager [req-0c1e76ea-32da-4eaf-a72a-a1287bc9c909 req-35b66b9f-82e6-4eea-b875-f620af015da0 service nova] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Received unexpected event network-vif-plugged-ef0eb4ad-a83b-435a-a813-4434a8616c14 for instance with vm_state building and task_state spawning. [ 847.963262] env[62753]: DEBUG nova.network.neutron [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Successfully updated port: ef0eb4ad-a83b-435a-a813-4434a8616c14 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 848.015869] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332166, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071354} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.015869] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 848.016552] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-207293d1-7fa7-4772-940e-05e610ab9ad6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.045043] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Reconfiguring VM instance instance-00000039 to attach disk [datastore2] 3bc05fdb-81be-4764-b9d9-01acd125b020/3bc05fdb-81be-4764-b9d9-01acd125b020.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 848.046307] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0d10a531-bd24-4578-ab08-aabd570fa3f8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.060549] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fd585c0-6f41-4c5f-ae7b-4da59fde510c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.067579] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "6b9a08b0-3cb3-474a-8193-953e74252a53" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 848.067861] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "6b9a08b0-3cb3-474a-8193-953e74252a53" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 848.073059] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1cc1ab05-d459-447e-8d91-7aef92f7a0f7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.076115] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 848.076115] env[62753]: value = "task-1332167" [ 848.076115] env[62753]: _type = "Task" [ 848.076115] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.104303] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-600f8584-4aa6-46d5-a11f-60b2b9f5509e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.109703] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332167, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.114686] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9dc619b5-0081-43bf-8339-80b4db9b85dd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.127727] env[62753]: DEBUG nova.compute.provider_tree [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 848.472092] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "refresh_cache-046fc3dc-a187-4fda-bc66-345e1226f83d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 848.472092] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired lock "refresh_cache-046fc3dc-a187-4fda-bc66-345e1226f83d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 848.472092] env[62753]: DEBUG nova.network.neutron [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 848.586099] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332167, 'name': ReconfigVM_Task, 'duration_secs': 0.275359} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 848.586099] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Reconfigured VM instance instance-00000039 to attach disk [datastore2] 3bc05fdb-81be-4764-b9d9-01acd125b020/3bc05fdb-81be-4764-b9d9-01acd125b020.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 848.586737] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-76eb3771-c434-4140-8d33-1d407d6862cb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 848.592946] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 848.592946] env[62753]: value = "task-1332168" [ 848.592946] env[62753]: _type = "Task" [ 848.592946] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 848.604410] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332168, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 848.631165] env[62753]: DEBUG nova.scheduler.client.report [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 849.005948] env[62753]: DEBUG nova.network.neutron [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 849.102878] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332168, 'name': Rename_Task, 'duration_secs': 0.13033} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.103152] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 849.103386] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-867bd2a3-7ba3-4db1-9119-29b196a66f6b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.109230] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 849.109230] env[62753]: value = "task-1332169" [ 849.109230] env[62753]: _type = "Task" [ 849.109230] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.117546] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332169, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.135737] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.927s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 849.136272] env[62753]: ERROR nova.compute.manager [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4aef7cff-9e33-4256-9f64-2960ecceb14b, please check neutron logs for more information. [ 849.136272] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Traceback (most recent call last): [ 849.136272] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 849.136272] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] self.driver.spawn(context, instance, image_meta, [ 849.136272] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 849.136272] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] self._vmops.spawn(context, instance, image_meta, injected_files, [ 849.136272] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 849.136272] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] vm_ref = self.build_virtual_machine(instance, [ 849.136272] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 849.136272] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] vif_infos = vmwarevif.get_vif_info(self._session, [ 849.136272] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 849.136656] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] for vif in network_info: [ 849.136656] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 849.136656] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] return self._sync_wrapper(fn, *args, **kwargs) [ 849.136656] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 849.136656] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] self.wait() [ 849.136656] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 849.136656] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] self[:] = self._gt.wait() [ 849.136656] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 849.136656] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] return self._exit_event.wait() [ 849.136656] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 849.136656] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] current.throw(*self._exc) [ 849.136656] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 849.136656] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] result = function(*args, **kwargs) [ 849.137019] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 849.137019] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] return func(*args, **kwargs) [ 849.137019] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 849.137019] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] raise e [ 849.137019] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 849.137019] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] nwinfo = self.network_api.allocate_for_instance( [ 849.137019] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 849.137019] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] created_port_ids = self._update_ports_for_instance( [ 849.137019] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 849.137019] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] with excutils.save_and_reraise_exception(): [ 849.137019] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 849.137019] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] self.force_reraise() [ 849.137019] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 849.137390] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] raise self.value [ 849.137390] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 849.137390] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] updated_port = self._update_port( [ 849.137390] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 849.137390] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] _ensure_no_port_binding_failure(port) [ 849.137390] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 849.137390] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] raise exception.PortBindingFailed(port_id=port['id']) [ 849.137390] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] nova.exception.PortBindingFailed: Binding failed for port 4aef7cff-9e33-4256-9f64-2960ecceb14b, please check neutron logs for more information. [ 849.137390] env[62753]: ERROR nova.compute.manager [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] [ 849.137390] env[62753]: DEBUG nova.compute.utils [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Binding failed for port 4aef7cff-9e33-4256-9f64-2960ecceb14b, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 849.138785] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.497s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 849.141630] env[62753]: DEBUG nova.compute.manager [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Build of instance 92e86d6a-e455-4984-9fdb-d1cdf856e729 was re-scheduled: Binding failed for port 4aef7cff-9e33-4256-9f64-2960ecceb14b, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 849.142111] env[62753]: DEBUG nova.compute.manager [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 849.142358] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Acquiring lock "refresh_cache-92e86d6a-e455-4984-9fdb-d1cdf856e729" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.142547] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Acquired lock "refresh_cache-92e86d6a-e455-4984-9fdb-d1cdf856e729" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.142719] env[62753]: DEBUG nova.network.neutron [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 849.151934] env[62753]: DEBUG nova.network.neutron [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Updating instance_info_cache with network_info: [{"id": "ef0eb4ad-a83b-435a-a813-4434a8616c14", "address": "fa:16:3e:aa:48:3c", "network": {"id": "8c5edc3b-7faf-47f2-9736-7093f9ff72de", "bridge": "br-int", "label": "tempest-ImagesTestJSON-588052284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e376836411ae4854965adf2923eaaffd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76f377cd-5966-49b4-9210-907f592c694e", "external-id": "nsx-vlan-transportzone-124", "segmentation_id": 124, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef0eb4ad-a8", "ovs_interfaceid": "ef0eb4ad-a83b-435a-a813-4434a8616c14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.619442] env[62753]: DEBUG oslo_vmware.api [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332169, 'name': PowerOnVM_Task, 'duration_secs': 0.442264} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 849.619722] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 849.619894] env[62753]: INFO nova.compute.manager [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Took 7.46 seconds to spawn the instance on the hypervisor. [ 849.620078] env[62753]: DEBUG nova.compute.manager [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 849.620859] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51f6aaea-6922-4a1c-a221-3840cc98fd30 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.653598] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Releasing lock "refresh_cache-046fc3dc-a187-4fda-bc66-345e1226f83d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 849.653989] env[62753]: DEBUG nova.compute.manager [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Instance network_info: |[{"id": "ef0eb4ad-a83b-435a-a813-4434a8616c14", "address": "fa:16:3e:aa:48:3c", "network": {"id": "8c5edc3b-7faf-47f2-9736-7093f9ff72de", "bridge": "br-int", "label": "tempest-ImagesTestJSON-588052284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e376836411ae4854965adf2923eaaffd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76f377cd-5966-49b4-9210-907f592c694e", "external-id": "nsx-vlan-transportzone-124", "segmentation_id": 124, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef0eb4ad-a8", "ovs_interfaceid": "ef0eb4ad-a83b-435a-a813-4434a8616c14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 849.654486] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:aa:48:3c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '76f377cd-5966-49b4-9210-907f592c694e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef0eb4ad-a83b-435a-a813-4434a8616c14', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 849.662583] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Creating folder: Project (e376836411ae4854965adf2923eaaffd). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.663662] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ac9b1883-e7e8-4b9c-8968-1bc694da5c2e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.667808] env[62753]: DEBUG nova.network.neutron [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 849.675114] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Created folder: Project (e376836411ae4854965adf2923eaaffd) in parent group-v284541. [ 849.675114] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Creating folder: Instances. Parent ref: group-v284564. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 849.675114] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-769e24e4-93aa-4017-b233-38c5702db612 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.684380] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Created folder: Instances in parent group-v284564. [ 849.684624] env[62753]: DEBUG oslo.service.loopingcall [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 849.687456] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 849.687846] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-2ec0e0c3-e716-4119-bb6d-101bd43879be {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.710752] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 849.710752] env[62753]: value = "task-1332172" [ 849.710752] env[62753]: _type = "Task" [ 849.710752] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 849.720690] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332172, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 849.794324] env[62753]: DEBUG nova.network.neutron [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 849.889230] env[62753]: DEBUG nova.compute.manager [req-77e97125-7bb0-4c86-924d-d7bb0a6f7f73 req-fd034dbf-3ede-46c6-975e-77a61a5b1edd service nova] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Received event network-changed-ef0eb4ad-a83b-435a-a813-4434a8616c14 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 849.889230] env[62753]: DEBUG nova.compute.manager [req-77e97125-7bb0-4c86-924d-d7bb0a6f7f73 req-fd034dbf-3ede-46c6-975e-77a61a5b1edd service nova] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Refreshing instance network info cache due to event network-changed-ef0eb4ad-a83b-435a-a813-4434a8616c14. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 849.889364] env[62753]: DEBUG oslo_concurrency.lockutils [req-77e97125-7bb0-4c86-924d-d7bb0a6f7f73 req-fd034dbf-3ede-46c6-975e-77a61a5b1edd service nova] Acquiring lock "refresh_cache-046fc3dc-a187-4fda-bc66-345e1226f83d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 849.889505] env[62753]: DEBUG oslo_concurrency.lockutils [req-77e97125-7bb0-4c86-924d-d7bb0a6f7f73 req-fd034dbf-3ede-46c6-975e-77a61a5b1edd service nova] Acquired lock "refresh_cache-046fc3dc-a187-4fda-bc66-345e1226f83d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 849.889663] env[62753]: DEBUG nova.network.neutron [req-77e97125-7bb0-4c86-924d-d7bb0a6f7f73 req-fd034dbf-3ede-46c6-975e-77a61a5b1edd service nova] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Refreshing network info cache for port ef0eb4ad-a83b-435a-a813-4434a8616c14 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 849.934348] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc47fa5c-afa1-45f5-bd2d-b3ffbaad0478 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.942587] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-42836e32-0a45-4d95-bfb0-5cf595d588a2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.975296] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-369e5dc9-3d13-4a7e-855a-cb8c5a397220 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.982943] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4fd0b324-bb93-4adf-af90-bdae4057e638 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 849.996390] env[62753]: DEBUG nova.compute.provider_tree [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 850.140381] env[62753]: INFO nova.compute.manager [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Took 28.07 seconds to build instance. [ 850.224789] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332172, 'name': CreateVM_Task, 'duration_secs': 0.340358} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.224789] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 850.224789] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.224789] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.224789] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 850.224789] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-49c8cbfb-af6a-454e-89a6-a3d18297dbc5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.230025] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 850.230025] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525d329e-2f16-e16f-9ea5-acbb66e88f88" [ 850.230025] env[62753]: _type = "Task" [ 850.230025] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.235714] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525d329e-2f16-e16f-9ea5-acbb66e88f88, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.299400] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Releasing lock "refresh_cache-92e86d6a-e455-4984-9fdb-d1cdf856e729" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.302867] env[62753]: DEBUG nova.compute.manager [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 850.302867] env[62753]: DEBUG nova.compute.manager [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 850.302867] env[62753]: DEBUG nova.network.neutron [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 850.322742] env[62753]: DEBUG nova.network.neutron [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 850.503020] env[62753]: DEBUG nova.scheduler.client.report [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 850.643446] env[62753]: DEBUG oslo_concurrency.lockutils [None req-310118ff-3f44-42ed-8997-46296a9e7f58 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "3bc05fdb-81be-4764-b9d9-01acd125b020" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 111.492s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 850.698067] env[62753]: DEBUG nova.network.neutron [req-77e97125-7bb0-4c86-924d-d7bb0a6f7f73 req-fd034dbf-3ede-46c6-975e-77a61a5b1edd service nova] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Updated VIF entry in instance network info cache for port ef0eb4ad-a83b-435a-a813-4434a8616c14. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 850.698067] env[62753]: DEBUG nova.network.neutron [req-77e97125-7bb0-4c86-924d-d7bb0a6f7f73 req-fd034dbf-3ede-46c6-975e-77a61a5b1edd service nova] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Updating instance_info_cache with network_info: [{"id": "ef0eb4ad-a83b-435a-a813-4434a8616c14", "address": "fa:16:3e:aa:48:3c", "network": {"id": "8c5edc3b-7faf-47f2-9736-7093f9ff72de", "bridge": "br-int", "label": "tempest-ImagesTestJSON-588052284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e376836411ae4854965adf2923eaaffd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76f377cd-5966-49b4-9210-907f592c694e", "external-id": "nsx-vlan-transportzone-124", "segmentation_id": 124, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef0eb4ad-a8", "ovs_interfaceid": "ef0eb4ad-a83b-435a-a813-4434a8616c14", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.738866] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525d329e-2f16-e16f-9ea5-acbb66e88f88, 'name': SearchDatastore_Task, 'duration_secs': 0.008686} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 850.739346] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 850.739712] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 850.740101] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.740382] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.740706] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 850.741349] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-79286d8a-ea2e-492b-8331-13212bc4acf7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.749340] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 850.752015] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 850.752015] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1cf12ba0-5b44-40e1-8bd5-c78a172bf1b3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 850.757131] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 850.757131] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523895a0-68cf-3d06-c859-5394948705da" [ 850.757131] env[62753]: _type = "Task" [ 850.757131] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 850.764661] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523895a0-68cf-3d06-c859-5394948705da, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 850.826687] env[62753]: DEBUG nova.network.neutron [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 850.862758] env[62753]: DEBUG nova.compute.manager [req-b53ba528-9440-4996-b2c5-c8bbf1d99a19 req-ffca2e73-d1c6-4136-bd8d-6d26ae35e7c4 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Received event network-changed-7e8e502a-7a64-4dfd-8f9d-18965ac19b6b {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 850.862952] env[62753]: DEBUG nova.compute.manager [req-b53ba528-9440-4996-b2c5-c8bbf1d99a19 req-ffca2e73-d1c6-4136-bd8d-6d26ae35e7c4 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Refreshing instance network info cache due to event network-changed-7e8e502a-7a64-4dfd-8f9d-18965ac19b6b. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 850.863245] env[62753]: DEBUG oslo_concurrency.lockutils [req-b53ba528-9440-4996-b2c5-c8bbf1d99a19 req-ffca2e73-d1c6-4136-bd8d-6d26ae35e7c4 service nova] Acquiring lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 850.863321] env[62753]: DEBUG oslo_concurrency.lockutils [req-b53ba528-9440-4996-b2c5-c8bbf1d99a19 req-ffca2e73-d1c6-4136-bd8d-6d26ae35e7c4 service nova] Acquired lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 850.863552] env[62753]: DEBUG nova.network.neutron [req-b53ba528-9440-4996-b2c5-c8bbf1d99a19 req-ffca2e73-d1c6-4136-bd8d-6d26ae35e7c4 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Refreshing network info cache for port 7e8e502a-7a64-4dfd-8f9d-18965ac19b6b {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 851.006627] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.867s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 851.006627] env[62753]: ERROR nova.compute.manager [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 60dffaa7-7a7a-4d14-88e2-fc60b0015fd0, please check neutron logs for more information. [ 851.006627] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Traceback (most recent call last): [ 851.006627] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 851.006627] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] self.driver.spawn(context, instance, image_meta, [ 851.006627] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 851.006627] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] self._vmops.spawn(context, instance, image_meta, injected_files, [ 851.006627] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 851.006627] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] vm_ref = self.build_virtual_machine(instance, [ 851.006865] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 851.006865] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] vif_infos = vmwarevif.get_vif_info(self._session, [ 851.006865] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 851.006865] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] for vif in network_info: [ 851.006865] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 851.006865] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] return self._sync_wrapper(fn, *args, **kwargs) [ 851.006865] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 851.006865] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] self.wait() [ 851.006865] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 851.006865] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] self[:] = self._gt.wait() [ 851.006865] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 851.006865] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] return self._exit_event.wait() [ 851.006865] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait [ 851.007259] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] result = hub.switch() [ 851.007259] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch [ 851.007259] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] return self.greenlet.switch() [ 851.007259] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 851.007259] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] result = function(*args, **kwargs) [ 851.007259] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 851.007259] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] return func(*args, **kwargs) [ 851.007259] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 851.007259] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] raise e [ 851.007259] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 851.007259] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] nwinfo = self.network_api.allocate_for_instance( [ 851.007259] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 851.007259] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] created_port_ids = self._update_ports_for_instance( [ 851.007531] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 851.007531] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] with excutils.save_and_reraise_exception(): [ 851.007531] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 851.007531] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] self.force_reraise() [ 851.007531] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 851.007531] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] raise self.value [ 851.007531] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 851.007531] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] updated_port = self._update_port( [ 851.007531] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 851.007531] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] _ensure_no_port_binding_failure(port) [ 851.007531] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 851.007531] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] raise exception.PortBindingFailed(port_id=port['id']) [ 851.007843] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] nova.exception.PortBindingFailed: Binding failed for port 60dffaa7-7a7a-4d14-88e2-fc60b0015fd0, please check neutron logs for more information. [ 851.007843] env[62753]: ERROR nova.compute.manager [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] [ 851.007843] env[62753]: DEBUG nova.compute.utils [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Binding failed for port 60dffaa7-7a7a-4d14-88e2-fc60b0015fd0, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 851.008636] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.032s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 851.010471] env[62753]: INFO nova.compute.claims [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 851.013270] env[62753]: DEBUG nova.compute.manager [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Build of instance 7931ae4b-c486-4e57-9bf5-14166a1d4201 was re-scheduled: Binding failed for port 60dffaa7-7a7a-4d14-88e2-fc60b0015fd0, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 851.013548] env[62753]: DEBUG nova.compute.manager [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 851.013759] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Acquiring lock "refresh_cache-7931ae4b-c486-4e57-9bf5-14166a1d4201" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 851.013908] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Acquired lock "refresh_cache-7931ae4b-c486-4e57-9bf5-14166a1d4201" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 851.014066] env[62753]: DEBUG nova.network.neutron [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 851.145844] env[62753]: DEBUG nova.compute.manager [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 851.200429] env[62753]: DEBUG oslo_concurrency.lockutils [req-77e97125-7bb0-4c86-924d-d7bb0a6f7f73 req-fd034dbf-3ede-46c6-975e-77a61a5b1edd service nova] Releasing lock "refresh_cache-046fc3dc-a187-4fda-bc66-345e1226f83d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.268664] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523895a0-68cf-3d06-c859-5394948705da, 'name': SearchDatastore_Task, 'duration_secs': 0.007508} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.269884] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f5e6ec46-2c05-4ea3-86bb-155502cf517a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.275051] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 851.275051] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5294a946-04b3-b3e6-5d4a-715fd75e6bd8" [ 851.275051] env[62753]: _type = "Task" [ 851.275051] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.286151] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5294a946-04b3-b3e6-5d4a-715fd75e6bd8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 851.328562] env[62753]: INFO nova.compute.manager [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] [instance: 92e86d6a-e455-4984-9fdb-d1cdf856e729] Took 1.03 seconds to deallocate network for instance. [ 851.537641] env[62753]: DEBUG nova.network.neutron [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 851.642019] env[62753]: DEBUG nova.network.neutron [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.673914] env[62753]: DEBUG oslo_concurrency.lockutils [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 851.728588] env[62753]: DEBUG nova.network.neutron [req-b53ba528-9440-4996-b2c5-c8bbf1d99a19 req-ffca2e73-d1c6-4136-bd8d-6d26ae35e7c4 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Updated VIF entry in instance network info cache for port 7e8e502a-7a64-4dfd-8f9d-18965ac19b6b. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 851.728588] env[62753]: DEBUG nova.network.neutron [req-b53ba528-9440-4996-b2c5-c8bbf1d99a19 req-ffca2e73-d1c6-4136-bd8d-6d26ae35e7c4 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Updating instance_info_cache with network_info: [{"id": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "address": "fa:16:3e:1d:84:1b", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8e502a-7a", "ovs_interfaceid": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 851.789941] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5294a946-04b3-b3e6-5d4a-715fd75e6bd8, 'name': SearchDatastore_Task, 'duration_secs': 0.008816} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 851.789941] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 851.789941] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 046fc3dc-a187-4fda-bc66-345e1226f83d/046fc3dc-a187-4fda-bc66-345e1226f83d.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 851.789941] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-939d06d0-661e-429f-8e3a-edb87d09d6ff {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 851.799955] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 851.799955] env[62753]: value = "task-1332173" [ 851.799955] env[62753]: _type = "Task" [ 851.799955] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 851.811552] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332173, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.149736] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Releasing lock "refresh_cache-7931ae4b-c486-4e57-9bf5-14166a1d4201" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.149736] env[62753]: DEBUG nova.compute.manager [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 852.149736] env[62753]: DEBUG nova.compute.manager [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 852.149972] env[62753]: DEBUG nova.network.neutron [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 852.171240] env[62753]: DEBUG nova.network.neutron [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 852.232166] env[62753]: DEBUG oslo_concurrency.lockutils [req-b53ba528-9440-4996-b2c5-c8bbf1d99a19 req-ffca2e73-d1c6-4136-bd8d-6d26ae35e7c4 service nova] Releasing lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 852.314080] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332173, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.494868} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.318530] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 046fc3dc-a187-4fda-bc66-345e1226f83d/046fc3dc-a187-4fda-bc66-345e1226f83d.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 852.318530] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 852.318761] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-66483150-4974-4b04-8c2c-9817f96e2a37 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.327753] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 852.327753] env[62753]: value = "task-1332174" [ 852.327753] env[62753]: _type = "Task" [ 852.327753] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.339237] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332174, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.348746] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7de7b720-c57d-4c20-95f8-01d9d9cd1688 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.357133] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7146ea09-5721-4154-8ed1-af30df2310d9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.362338] env[62753]: INFO nova.scheduler.client.report [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Deleted allocations for instance 92e86d6a-e455-4984-9fdb-d1cdf856e729 [ 852.395687] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f3a2d4d-fac4-435c-a09b-f64dc93f9584 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.403446] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6bdc0bc3-753a-44e8-b170-8724ac0468a1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.416877] env[62753]: DEBUG nova.compute.provider_tree [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 852.674690] env[62753]: DEBUG nova.network.neutron [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 852.836750] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332174, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.066299} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 852.837013] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 852.837892] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3be12501-c2f8-4eeb-84b7-b282bccec7fe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.861508] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Reconfiguring VM instance instance-0000003a to attach disk [datastore2] 046fc3dc-a187-4fda-bc66-345e1226f83d/046fc3dc-a187-4fda-bc66-345e1226f83d.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 852.861628] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12a7aba8-ed80-4dda-98ad-9fd5a61a10e6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 852.884952] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 852.884952] env[62753]: value = "task-1332175" [ 852.884952] env[62753]: _type = "Task" [ 852.884952] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 852.893590] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332175, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 852.899145] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a14a551c-8fa9-4dac-8c44-fed6d490f5bf tempest-ServerActionsTestOtherA-1193232944 tempest-ServerActionsTestOtherA-1193232944-project-member] Lock "92e86d6a-e455-4984-9fdb-d1cdf856e729" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 156.621s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 852.921138] env[62753]: DEBUG nova.scheduler.client.report [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 853.178161] env[62753]: INFO nova.compute.manager [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] [instance: 7931ae4b-c486-4e57-9bf5-14166a1d4201] Took 1.03 seconds to deallocate network for instance. [ 853.395263] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332175, 'name': ReconfigVM_Task, 'duration_secs': 0.28126} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.395649] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Reconfigured VM instance instance-0000003a to attach disk [datastore2] 046fc3dc-a187-4fda-bc66-345e1226f83d/046fc3dc-a187-4fda-bc66-345e1226f83d.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 853.396299] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-92a8175b-f236-44d1-affb-c7afd604bff8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.401690] env[62753]: DEBUG nova.compute.manager [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 853.404983] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 853.404983] env[62753]: value = "task-1332176" [ 853.404983] env[62753]: _type = "Task" [ 853.404983] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.416866] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332176, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 853.426480] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.418s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 853.426993] env[62753]: DEBUG nova.compute.manager [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 853.430568] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: waited 18.149s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 853.921691] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332176, 'name': Rename_Task, 'duration_secs': 0.132468} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 853.921999] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 853.921999] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-54ae9f3b-9e4a-467f-8873-99bc9b8ee4b8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 853.928504] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 853.928504] env[62753]: value = "task-1332177" [ 853.928504] env[62753]: _type = "Task" [ 853.928504] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 853.933951] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 853.935264] env[62753]: DEBUG nova.compute.utils [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 853.940326] env[62753]: DEBUG nova.compute.manager [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Not allocating networking since 'none' was specified. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 853.948256] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332177, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 854.190139] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd51760e-085e-403c-ad90-931c1c0eb69d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.198229] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61f96b1c-e112-498c-bfa8-36e969df708c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.230029] env[62753]: INFO nova.scheduler.client.report [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Deleted allocations for instance 7931ae4b-c486-4e57-9bf5-14166a1d4201 [ 854.235993] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd86a9f0-6f94-46a3-a1d8-869d858472fa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.245093] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69016922-10a1-4bf0-a4b3-a6ac3cc2bd93 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.261248] env[62753]: DEBUG nova.compute.provider_tree [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 854.440188] env[62753]: DEBUG oslo_vmware.api [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332177, 'name': PowerOnVM_Task, 'duration_secs': 0.466344} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 854.440188] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 854.440188] env[62753]: INFO nova.compute.manager [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Took 7.12 seconds to spawn the instance on the hypervisor. [ 854.440188] env[62753]: DEBUG nova.compute.manager [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 854.440321] env[62753]: DEBUG nova.compute.manager [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 854.443223] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f210fe9-ebc9-4e85-b062-7d80b6949a6d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 854.450688] env[62753]: DEBUG oslo_concurrency.lockutils [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "c73b1ae0-dc98-47f7-babf-e96169384785" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 854.450940] env[62753]: DEBUG oslo_concurrency.lockutils [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "c73b1ae0-dc98-47f7-babf-e96169384785" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 854.741851] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1c75f7b9-e685-4977-8a33-cdc0d390870c tempest-SecurityGroupsTestJSON-435262203 tempest-SecurityGroupsTestJSON-435262203-project-member] Lock "7931ae4b-c486-4e57-9bf5-14166a1d4201" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 126.777s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 854.765052] env[62753]: DEBUG nova.scheduler.client.report [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 854.966908] env[62753]: INFO nova.compute.manager [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Took 28.60 seconds to build instance. [ 855.245742] env[62753]: DEBUG nova.compute.manager [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 855.273387] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.abort_instance_claim" :: held 1.840s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.273387] env[62753]: ERROR nova.compute.manager [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Failed to build and run instance: nova.exception.PortBindingFailed: Binding failed for port 4a43717d-e663-4998-a93a-fc8c773710d3, please check neutron logs for more information. [ 855.273387] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] Traceback (most recent call last): [ 855.273387] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/compute/manager.py", line 2641, in _build_and_run_instance [ 855.273387] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] self.driver.spawn(context, instance, image_meta, [ 855.273387] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 546, in spawn [ 855.273387] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] self._vmops.spawn(context, instance, image_meta, injected_files, [ 855.273387] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 753, in spawn [ 855.273387] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] vm_ref = self.build_virtual_machine(instance, [ 855.273681] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/virt/vmwareapi/vmops.py", line 275, in build_virtual_machine [ 855.273681] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] vif_infos = vmwarevif.get_vif_info(self._session, [ 855.273681] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/virt/vmwareapi/vif.py", line 119, in get_vif_info [ 855.273681] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] for vif in network_info: [ 855.273681] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/network/model.py", line 612, in __iter__ [ 855.273681] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] return self._sync_wrapper(fn, *args, **kwargs) [ 855.273681] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/network/model.py", line 603, in _sync_wrapper [ 855.273681] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] self.wait() [ 855.273681] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/network/model.py", line 635, in wait [ 855.273681] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] self[:] = self._gt.wait() [ 855.273681] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 225, in wait [ 855.273681] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] return self._exit_event.wait() [ 855.273681] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 131, in wait [ 855.273968] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] current.throw(*self._exc) [ 855.273968] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/greenthread.py", line 265, in main [ 855.273968] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] result = function(*args, **kwargs) [ 855.273968] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/utils.py", line 663, in context_wrapper [ 855.273968] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] return func(*args, **kwargs) [ 855.273968] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/compute/manager.py", line 2011, in _allocate_network_async [ 855.273968] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] raise e [ 855.273968] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/compute/manager.py", line 1989, in _allocate_network_async [ 855.273968] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] nwinfo = self.network_api.allocate_for_instance( [ 855.273968] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/network/neutron.py", line 1278, in allocate_for_instance [ 855.273968] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] created_port_ids = self._update_ports_for_instance( [ 855.273968] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/network/neutron.py", line 1414, in _update_ports_for_instance [ 855.273968] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] with excutils.save_and_reraise_exception(): [ 855.274267] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 855.274267] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] self.force_reraise() [ 855.274267] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 855.274267] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] raise self.value [ 855.274267] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/network/neutron.py", line 1389, in _update_ports_for_instance [ 855.274267] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] updated_port = self._update_port( [ 855.274267] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/network/neutron.py", line 585, in _update_port [ 855.274267] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] _ensure_no_port_binding_failure(port) [ 855.274267] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] File "/opt/stack/nova/nova/network/neutron.py", line 294, in _ensure_no_port_binding_failure [ 855.274267] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] raise exception.PortBindingFailed(port_id=port['id']) [ 855.274267] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] nova.exception.PortBindingFailed: Binding failed for port 4a43717d-e663-4998-a93a-fc8c773710d3, please check neutron logs for more information. [ 855.274267] env[62753]: ERROR nova.compute.manager [instance: 16c82d55-1420-4f83-a547-295e793de9df] [ 855.274552] env[62753]: DEBUG nova.compute.utils [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Binding failed for port 4a43717d-e663-4998-a93a-fc8c773710d3, please check neutron logs for more information. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 855.274552] env[62753]: DEBUG oslo_concurrency.lockutils [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.256s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 855.276608] env[62753]: INFO nova.compute.claims [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 855.279348] env[62753]: DEBUG nova.compute.manager [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Build of instance 16c82d55-1420-4f83-a547-295e793de9df was re-scheduled: Binding failed for port 4a43717d-e663-4998-a93a-fc8c773710d3, please check neutron logs for more information. {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2480}} [ 855.279472] env[62753]: DEBUG nova.compute.manager [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Unplugging VIFs for instance {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3006}} [ 855.279710] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "refresh_cache-16c82d55-1420-4f83-a547-295e793de9df" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 855.279857] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquired lock "refresh_cache-16c82d55-1420-4f83-a547-295e793de9df" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 855.280038] env[62753]: DEBUG nova.network.neutron [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 855.453406] env[62753]: DEBUG nova.compute.manager [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 855.470418] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8b4ff868-9a43-4047-90a6-1610848a364e tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "046fc3dc-a187-4fda-bc66-345e1226f83d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 102.494s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 855.486840] env[62753]: DEBUG nova.virt.hardware [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 855.487103] env[62753]: DEBUG nova.virt.hardware [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 855.487263] env[62753]: DEBUG nova.virt.hardware [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 855.487444] env[62753]: DEBUG nova.virt.hardware [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 855.487591] env[62753]: DEBUG nova.virt.hardware [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 855.487733] env[62753]: DEBUG nova.virt.hardware [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 855.487939] env[62753]: DEBUG nova.virt.hardware [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 855.488113] env[62753]: DEBUG nova.virt.hardware [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 855.488281] env[62753]: DEBUG nova.virt.hardware [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 855.488443] env[62753]: DEBUG nova.virt.hardware [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 855.488612] env[62753]: DEBUG nova.virt.hardware [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 855.489820] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18171699-a924-4cd4-a8b4-f30b6db05457 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.498964] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6099ad66-f7f0-4463-ae87-46a1084f60e8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.513042] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Instance VIF info [] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 855.519080] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Creating folder: Project (98fb09d6c4f349d68ee12c534a822de2). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 855.519799] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1669c9b-1bbc-4aac-8463-82d7530278e5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.533176] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Created folder: Project (98fb09d6c4f349d68ee12c534a822de2) in parent group-v284541. [ 855.533176] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Creating folder: Instances. Parent ref: group-v284567. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 855.533176] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-230820c7-48e7-443e-88d3-e4224192500d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.541800] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Created folder: Instances in parent group-v284567. [ 855.542075] env[62753]: DEBUG oslo.service.loopingcall [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 855.542303] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 855.542592] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b4b18e3c-af68-4dd7-8cf3-414ef8907490 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 855.564923] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 855.564923] env[62753]: value = "task-1332180" [ 855.564923] env[62753]: _type = "Task" [ 855.564923] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 855.573584] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332180, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 855.779178] env[62753]: DEBUG oslo_concurrency.lockutils [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 855.811974] env[62753]: DEBUG nova.network.neutron [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 855.934207] env[62753]: DEBUG nova.network.neutron [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 855.973127] env[62753]: DEBUG nova.compute.manager [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 856.074968] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332180, 'name': CreateVM_Task, 'duration_secs': 0.32559} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.075365] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 856.075949] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.077081] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.077081] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 856.077081] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-52e23cfa-48f1-4184-b0ad-09c51a7c1e1b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.082845] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 856.082845] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bae1f7-1fe5-8540-df4b-1eb504003cef" [ 856.082845] env[62753]: _type = "Task" [ 856.082845] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.091368] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bae1f7-1fe5-8540-df4b-1eb504003cef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.353083] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae302b5-a8a6-4a4c-9957-758bbe235991 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.370388] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b8b5b286-a184-4c00-b1d4-69fd1df6b005 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Suspending the VM {{(pid=62753) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 856.371085] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-cbbdbabf-e01c-4445-8f9e-f5122d8aa3fa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.378139] env[62753]: DEBUG oslo_vmware.api [None req-b8b5b286-a184-4c00-b1d4-69fd1df6b005 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 856.378139] env[62753]: value = "task-1332181" [ 856.378139] env[62753]: _type = "Task" [ 856.378139] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.388460] env[62753]: DEBUG oslo_vmware.api [None req-b8b5b286-a184-4c00-b1d4-69fd1df6b005 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332181, 'name': SuspendVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.437556] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Releasing lock "refresh_cache-16c82d55-1420-4f83-a547-295e793de9df" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.437791] env[62753]: DEBUG nova.compute.manager [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Virt driver does not provide unplug_vifs method, so it is not possible determine if VIFs should be unplugged. {{(pid=62753) _cleanup_allocated_networks /opt/stack/nova/nova/compute/manager.py:3029}} [ 856.437970] env[62753]: DEBUG nova.compute.manager [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 856.438154] env[62753]: DEBUG nova.network.neutron [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 856.454941] env[62753]: DEBUG nova.network.neutron [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 856.491807] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 856.540301] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f01a4909-dded-4cc2-ab4f-d588e817cdca {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.548281] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b99fb51-b720-4fd0-b5dc-4d8808d878e4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.579433] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f315da2a-8db3-4a84-8cee-908fe93c1c8d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.594700] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bae1f7-1fe5-8540-df4b-1eb504003cef, 'name': SearchDatastore_Task, 'duration_secs': 0.010377} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 856.594700] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 856.594700] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 856.594935] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 856.595092] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 856.595265] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 856.596548] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ca52ef6-b18b-4482-b99d-a78541a41bad {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.600725] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b7cff7d0-93ef-49fa-8538-b3d612b0ca03 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.613948] env[62753]: DEBUG nova.compute.provider_tree [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 856.616474] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 856.616650] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 856.617597] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d4a72769-db85-4d86-b53b-e0a6834f0115 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 856.622754] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 856.622754] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]527220e8-0609-8123-188c-368627b393cf" [ 856.622754] env[62753]: _type = "Task" [ 856.622754] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 856.631020] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]527220e8-0609-8123-188c-368627b393cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.892327] env[62753]: DEBUG oslo_vmware.api [None req-b8b5b286-a184-4c00-b1d4-69fd1df6b005 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332181, 'name': SuspendVM_Task} progress is 62%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 856.958062] env[62753]: DEBUG nova.network.neutron [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 857.118767] env[62753]: DEBUG nova.scheduler.client.report [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 857.140801] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]527220e8-0609-8123-188c-368627b393cf, 'name': SearchDatastore_Task, 'duration_secs': 0.01974} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.142351] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8162efaa-2e02-4ee6-86dc-48b430ee73a3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.149311] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 857.149311] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]520aa058-34a5-1a32-2fb0-9d8f18eac38f" [ 857.149311] env[62753]: _type = "Task" [ 857.149311] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.157440] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]520aa058-34a5-1a32-2fb0-9d8f18eac38f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 857.393333] env[62753]: DEBUG oslo_vmware.api [None req-b8b5b286-a184-4c00-b1d4-69fd1df6b005 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332181, 'name': SuspendVM_Task, 'duration_secs': 0.636641} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.393602] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b8b5b286-a184-4c00-b1d4-69fd1df6b005 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Suspended the VM {{(pid=62753) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 857.393910] env[62753]: DEBUG nova.compute.manager [None req-b8b5b286-a184-4c00-b1d4-69fd1df6b005 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 857.394748] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-897b014d-56cc-404d-a080-5eba31531d85 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.461705] env[62753]: INFO nova.compute.manager [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 16c82d55-1420-4f83-a547-295e793de9df] Took 1.02 seconds to deallocate network for instance. [ 857.624191] env[62753]: DEBUG oslo_concurrency.lockutils [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.350s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 857.624745] env[62753]: DEBUG nova.compute.manager [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 857.631021] env[62753]: DEBUG oslo_concurrency.lockutils [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 857.631021] env[62753]: INFO nova.compute.claims [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 857.671868] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]520aa058-34a5-1a32-2fb0-9d8f18eac38f, 'name': SearchDatastore_Task, 'duration_secs': 0.009473} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 857.673607] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 857.674392] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3/4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 857.674392] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-1d1bb0cf-147d-422b-9fc6-12147c6962eb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 857.683745] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 857.683745] env[62753]: value = "task-1332182" [ 857.683745] env[62753]: _type = "Task" [ 857.683745] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 857.693730] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332182, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.131290] env[62753]: DEBUG nova.compute.utils [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 858.131681] env[62753]: DEBUG nova.compute.manager [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 858.135812] env[62753]: DEBUG nova.network.neutron [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 858.194108] env[62753]: DEBUG nova.policy [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9a00ada439d94a6a88c0393f8f1aff7a', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5056a6afa2f047dca12d9e32d5705c35', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 858.198362] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332182, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.464732] env[62753]: DEBUG nova.network.neutron [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Successfully created port: 5af88b4d-699d-4bac-b78f-f66b69bc1eea {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 858.507387] env[62753]: INFO nova.scheduler.client.report [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Deleted allocations for instance 16c82d55-1420-4f83-a547-295e793de9df [ 858.637743] env[62753]: DEBUG nova.compute.manager [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 858.695337] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332182, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.512805} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 858.698217] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3/4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 858.698478] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 858.698912] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fe6e670e-5c91-4491-ab3e-65272da3023e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.705799] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 858.705799] env[62753]: value = "task-1332183" [ 858.705799] env[62753]: _type = "Task" [ 858.705799] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 858.713654] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332183, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 858.877793] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db58e97a-ff83-47f9-b55f-9f86e1a76736 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.885787] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b2cd6b-db3a-4ff0-8f5c-a2ca6eadafea {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.916655] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1be925e9-deed-45e3-8d3a-6a4b3c60b25c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.923695] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a98fb0d4-72c4-46d6-b4c1-07c842877a93 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 858.939014] env[62753]: DEBUG nova.compute.provider_tree [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 859.017586] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a38a9cab-d790-4106-bcbe-6d1246f9c9e0 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "16c82d55-1420-4f83-a547-295e793de9df" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 120.068s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.215363] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332183, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065146} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.215653] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 859.216414] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d7339cd-8f48-47c7-a3fc-4775618806a4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.236260] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Reconfiguring VM instance instance-0000003b to attach disk [datastore2] 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3/4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 859.236527] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8e6744b0-0c2b-417d-8404-bddfd08ecb28 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.255902] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 859.255902] env[62753]: value = "task-1332184" [ 859.255902] env[62753]: _type = "Task" [ 859.255902] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.263890] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332184, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.442210] env[62753]: DEBUG nova.scheduler.client.report [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 859.520617] env[62753]: DEBUG nova.compute.manager [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 859.651566] env[62753]: DEBUG nova.compute.manager [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 859.686158] env[62753]: DEBUG nova.virt.hardware [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 859.686450] env[62753]: DEBUG nova.virt.hardware [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 859.686610] env[62753]: DEBUG nova.virt.hardware [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 859.686790] env[62753]: DEBUG nova.virt.hardware [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 859.686938] env[62753]: DEBUG nova.virt.hardware [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 859.687321] env[62753]: DEBUG nova.virt.hardware [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 859.687565] env[62753]: DEBUG nova.virt.hardware [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 859.687730] env[62753]: DEBUG nova.virt.hardware [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 859.687904] env[62753]: DEBUG nova.virt.hardware [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 859.688074] env[62753]: DEBUG nova.virt.hardware [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 859.688253] env[62753]: DEBUG nova.virt.hardware [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 859.689454] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbb8a3ec-c3fb-4f78-b3bb-2a9625dcc271 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.698239] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd760c7-02e9-4aa3-93bd-b6481c6cbb65 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.766178] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332184, 'name': ReconfigVM_Task, 'duration_secs': 0.269001} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 859.766490] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Reconfigured VM instance instance-0000003b to attach disk [datastore2] 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3/4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 859.767234] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a6340a4d-59c5-413b-980e-5051eca76255 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.774886] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 859.774886] env[62753]: value = "task-1332185" [ 859.774886] env[62753]: _type = "Task" [ 859.774886] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 859.783383] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332185, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 859.785633] env[62753]: DEBUG nova.compute.manager [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 859.786398] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12b3b955-acba-420d-9680-5d9202b393cd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 859.947202] env[62753]: DEBUG oslo_concurrency.lockutils [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.319s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 859.947725] env[62753]: DEBUG nova.compute.manager [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 859.950355] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 17.762s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 859.951823] env[62753]: INFO nova.compute.claims [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 860.005120] env[62753]: DEBUG nova.network.neutron [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Successfully updated port: 5af88b4d-699d-4bac-b78f-f66b69bc1eea {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 860.042624] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.067093] env[62753]: DEBUG nova.compute.manager [req-ef191e5e-4360-4c92-ad9a-d6386bc0d527 req-d41e4fc7-ebf3-4102-b476-881635a4b2f0 service nova] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Received event network-vif-plugged-5af88b4d-699d-4bac-b78f-f66b69bc1eea {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 860.067339] env[62753]: DEBUG oslo_concurrency.lockutils [req-ef191e5e-4360-4c92-ad9a-d6386bc0d527 req-d41e4fc7-ebf3-4102-b476-881635a4b2f0 service nova] Acquiring lock "f132230d-0fba-4293-b940-85cc7ef1cad1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 860.067937] env[62753]: DEBUG oslo_concurrency.lockutils [req-ef191e5e-4360-4c92-ad9a-d6386bc0d527 req-d41e4fc7-ebf3-4102-b476-881635a4b2f0 service nova] Lock "f132230d-0fba-4293-b940-85cc7ef1cad1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 860.068244] env[62753]: DEBUG oslo_concurrency.lockutils [req-ef191e5e-4360-4c92-ad9a-d6386bc0d527 req-d41e4fc7-ebf3-4102-b476-881635a4b2f0 service nova] Lock "f132230d-0fba-4293-b940-85cc7ef1cad1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 860.068432] env[62753]: DEBUG nova.compute.manager [req-ef191e5e-4360-4c92-ad9a-d6386bc0d527 req-d41e4fc7-ebf3-4102-b476-881635a4b2f0 service nova] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] No waiting events found dispatching network-vif-plugged-5af88b4d-699d-4bac-b78f-f66b69bc1eea {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 860.068866] env[62753]: WARNING nova.compute.manager [req-ef191e5e-4360-4c92-ad9a-d6386bc0d527 req-d41e4fc7-ebf3-4102-b476-881635a4b2f0 service nova] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Received unexpected event network-vif-plugged-5af88b4d-699d-4bac-b78f-f66b69bc1eea for instance with vm_state building and task_state spawning. [ 860.286522] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332185, 'name': Rename_Task, 'duration_secs': 0.129258} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.287714] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 860.287714] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-339310be-dcf8-4318-b065-30496c6f6c06 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.293293] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 860.293293] env[62753]: value = "task-1332186" [ 860.293293] env[62753]: _type = "Task" [ 860.293293] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.297204] env[62753]: INFO nova.compute.manager [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] instance snapshotting [ 860.297384] env[62753]: WARNING nova.compute.manager [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] trying to snapshot a non-running instance: (state: 7 expected: 1) [ 860.299914] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7132f3c8-37a5-42bb-bf3b-1ef2b1b2749b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.305157] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332186, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.323942] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-37481318-f4cd-42c2-a004-aab78ee43d56 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.456367] env[62753]: DEBUG nova.compute.utils [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 860.461318] env[62753]: DEBUG nova.compute.manager [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 860.461767] env[62753]: DEBUG nova.network.neutron [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 860.501346] env[62753]: DEBUG nova.policy [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b6207815e8041b4bba3069635d7d962', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6c8a6bc946a469fa85cc3ab80d7333a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 860.507362] env[62753]: DEBUG oslo_concurrency.lockutils [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Acquiring lock "refresh_cache-f132230d-0fba-4293-b940-85cc7ef1cad1" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 860.507528] env[62753]: DEBUG oslo_concurrency.lockutils [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Acquired lock "refresh_cache-f132230d-0fba-4293-b940-85cc7ef1cad1" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 860.507640] env[62753]: DEBUG nova.network.neutron [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 860.783066] env[62753]: DEBUG nova.network.neutron [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Successfully created port: bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 860.807991] env[62753]: DEBUG oslo_vmware.api [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332186, 'name': PowerOnVM_Task, 'duration_secs': 0.510595} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 860.807991] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 860.808358] env[62753]: INFO nova.compute.manager [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Took 5.35 seconds to spawn the instance on the hypervisor. [ 860.808604] env[62753]: DEBUG nova.compute.manager [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 860.809607] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-32312734-e1a6-480a-aac8-4af081c1af90 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.835441] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Creating Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 860.835904] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d41cbe4a-48dc-4a10-bc76-1f2c123603b4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 860.843757] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 860.843757] env[62753]: value = "task-1332187" [ 860.843757] env[62753]: _type = "Task" [ 860.843757] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 860.852746] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332187, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 860.962297] env[62753]: DEBUG nova.compute.manager [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 861.078087] env[62753]: DEBUG nova.network.neutron [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 861.330482] env[62753]: INFO nova.compute.manager [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Took 28.46 seconds to build instance. [ 861.342607] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e8c7d52d-8b83-4e4f-b7b5-2742784a7b5e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.361374] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-042a5cc7-a121-44ea-8bb7-0520730f5ecb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.364906] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332187, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 861.400644] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10406c34-e63f-42e2-9165-3285a65150db {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.409677] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-44386227-2d8a-460a-9f63-2a9f831ac220 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.430331] env[62753]: DEBUG nova.compute.provider_tree [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 861.536738] env[62753]: DEBUG nova.network.neutron [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Updating instance_info_cache with network_info: [{"id": "5af88b4d-699d-4bac-b78f-f66b69bc1eea", "address": "fa:16:3e:d9:d5:6d", "network": {"id": "8baa5dc5-b65f-49f4-a6f5-f6e5b66ddb17", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-941849868-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5056a6afa2f047dca12d9e32d5705c35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e238ac23-819b-452f-9015-52922e45efd3", "external-id": "nsx-vlan-transportzone-127", "segmentation_id": 127, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5af88b4d-69", "ovs_interfaceid": "5af88b4d-699d-4bac-b78f-f66b69bc1eea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 861.833076] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd4226c6-2948-4ab5-82ab-36f6d49b4a40 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Lock "4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.110s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 861.862118] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332187, 'name': CreateSnapshot_Task, 'duration_secs': 0.532414} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 861.862346] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Created Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 861.863160] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1124a9db-c2a8-4617-9f93-ae5da8fe8cfe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 861.918765] env[62753]: INFO nova.compute.manager [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Rebuilding instance [ 861.936804] env[62753]: DEBUG nova.scheduler.client.report [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 861.978153] env[62753]: DEBUG nova.compute.manager [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 861.983617] env[62753]: DEBUG nova.compute.manager [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 861.984686] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e92c087-c61f-4b92-85b3-88c81cc9c32e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.007690] env[62753]: DEBUG nova.virt.hardware [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 862.007937] env[62753]: DEBUG nova.virt.hardware [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 862.008101] env[62753]: DEBUG nova.virt.hardware [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 862.008358] env[62753]: DEBUG nova.virt.hardware [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 862.008447] env[62753]: DEBUG nova.virt.hardware [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 862.008553] env[62753]: DEBUG nova.virt.hardware [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 862.008754] env[62753]: DEBUG nova.virt.hardware [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 862.008906] env[62753]: DEBUG nova.virt.hardware [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 862.009076] env[62753]: DEBUG nova.virt.hardware [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 862.009234] env[62753]: DEBUG nova.virt.hardware [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 862.009400] env[62753]: DEBUG nova.virt.hardware [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 862.010254] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70a45a95-9c91-49a7-bac4-a1c97344ea1f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.017892] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b8a8617-ed04-4993-805c-36d0892a944e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.041806] env[62753]: DEBUG oslo_concurrency.lockutils [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Releasing lock "refresh_cache-f132230d-0fba-4293-b940-85cc7ef1cad1" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 862.042120] env[62753]: DEBUG nova.compute.manager [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Instance network_info: |[{"id": "5af88b4d-699d-4bac-b78f-f66b69bc1eea", "address": "fa:16:3e:d9:d5:6d", "network": {"id": "8baa5dc5-b65f-49f4-a6f5-f6e5b66ddb17", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-941849868-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5056a6afa2f047dca12d9e32d5705c35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e238ac23-819b-452f-9015-52922e45efd3", "external-id": "nsx-vlan-transportzone-127", "segmentation_id": 127, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5af88b4d-69", "ovs_interfaceid": "5af88b4d-699d-4bac-b78f-f66b69bc1eea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 862.042506] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d9:d5:6d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'e238ac23-819b-452f-9015-52922e45efd3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '5af88b4d-699d-4bac-b78f-f66b69bc1eea', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 862.050582] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Creating folder: Project (5056a6afa2f047dca12d9e32d5705c35). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 862.051164] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-97877628-f0d9-4d52-978c-521b9fb9c975 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.061101] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Created folder: Project (5056a6afa2f047dca12d9e32d5705c35) in parent group-v284541. [ 862.061301] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Creating folder: Instances. Parent ref: group-v284571. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 862.061549] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-7366311f-7ca0-4dc2-9a33-c83cfa968aca {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.071732] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Created folder: Instances in parent group-v284571. [ 862.071732] env[62753]: DEBUG oslo.service.loopingcall [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 862.071732] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 862.072118] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-aef45ee2-6aa7-4eac-b2d6-f1f23082cfb5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.090254] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 862.090254] env[62753]: value = "task-1332190" [ 862.090254] env[62753]: _type = "Task" [ 862.090254] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.096250] env[62753]: DEBUG nova.compute.manager [req-43116918-8399-48ef-a788-401e6b80e0c4 req-ae2029de-e6b1-4891-a70e-d2e4dcbf203c service nova] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Received event network-changed-5af88b4d-699d-4bac-b78f-f66b69bc1eea {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 862.096609] env[62753]: DEBUG nova.compute.manager [req-43116918-8399-48ef-a788-401e6b80e0c4 req-ae2029de-e6b1-4891-a70e-d2e4dcbf203c service nova] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Refreshing instance network info cache due to event network-changed-5af88b4d-699d-4bac-b78f-f66b69bc1eea. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 862.096976] env[62753]: DEBUG oslo_concurrency.lockutils [req-43116918-8399-48ef-a788-401e6b80e0c4 req-ae2029de-e6b1-4891-a70e-d2e4dcbf203c service nova] Acquiring lock "refresh_cache-f132230d-0fba-4293-b940-85cc7ef1cad1" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.097264] env[62753]: DEBUG oslo_concurrency.lockutils [req-43116918-8399-48ef-a788-401e6b80e0c4 req-ae2029de-e6b1-4891-a70e-d2e4dcbf203c service nova] Acquired lock "refresh_cache-f132230d-0fba-4293-b940-85cc7ef1cad1" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.097567] env[62753]: DEBUG nova.network.neutron [req-43116918-8399-48ef-a788-401e6b80e0c4 req-ae2029de-e6b1-4891-a70e-d2e4dcbf203c service nova] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Refreshing network info cache for port 5af88b4d-699d-4bac-b78f-f66b69bc1eea {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 862.106053] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332190, 'name': CreateVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.334826] env[62753]: DEBUG nova.compute.manager [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 862.340864] env[62753]: DEBUG nova.network.neutron [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Successfully updated port: bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 862.384579] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Creating linked-clone VM from snapshot {{(pid=62753) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 862.385263] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-6a1cb3b4-3831-43c5-9455-0b41d4b411f4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.393963] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 862.393963] env[62753]: value = "task-1332191" [ 862.393963] env[62753]: _type = "Task" [ 862.393963] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.406588] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332191, 'name': CloneVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.444338] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.494s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 862.445011] env[62753]: DEBUG nova.compute.manager [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 862.448214] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 19.435s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 862.499873] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 862.500179] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0035b400-c5bf-4ff8-bf11-37ab1e28f687 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.506701] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 862.506701] env[62753]: value = "task-1332192" [ 862.506701] env[62753]: _type = "Task" [ 862.506701] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.518109] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332192, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.602502] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332190, 'name': CreateVM_Task, 'duration_secs': 0.396558} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 862.602875] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 862.603923] env[62753]: DEBUG oslo_concurrency.lockutils [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.604218] env[62753]: DEBUG oslo_concurrency.lockutils [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.604639] env[62753]: DEBUG oslo_concurrency.lockutils [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 862.605025] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b426a04b-48b2-466c-bff7-40cdead52097 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 862.612019] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Waiting for the task: (returnval){ [ 862.612019] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52aa663b-e800-8093-59b8-27d943dedf8f" [ 862.612019] env[62753]: _type = "Task" [ 862.612019] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 862.623159] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52aa663b-e800-8093-59b8-27d943dedf8f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.843271] env[62753]: DEBUG oslo_concurrency.lockutils [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "refresh_cache-a6ee010c-6305-4009-80e0-92a2c58bcd7b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 862.843415] env[62753]: DEBUG oslo_concurrency.lockutils [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "refresh_cache-a6ee010c-6305-4009-80e0-92a2c58bcd7b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 862.843573] env[62753]: DEBUG nova.network.neutron [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 862.847855] env[62753]: DEBUG nova.network.neutron [req-43116918-8399-48ef-a788-401e6b80e0c4 req-ae2029de-e6b1-4891-a70e-d2e4dcbf203c service nova] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Updated VIF entry in instance network info cache for port 5af88b4d-699d-4bac-b78f-f66b69bc1eea. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 862.848202] env[62753]: DEBUG nova.network.neutron [req-43116918-8399-48ef-a788-401e6b80e0c4 req-ae2029de-e6b1-4891-a70e-d2e4dcbf203c service nova] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Updating instance_info_cache with network_info: [{"id": "5af88b4d-699d-4bac-b78f-f66b69bc1eea", "address": "fa:16:3e:d9:d5:6d", "network": {"id": "8baa5dc5-b65f-49f4-a6f5-f6e5b66ddb17", "bridge": "br-int", "label": "tempest-ServerTagsTestJSON-941849868-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5056a6afa2f047dca12d9e32d5705c35", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "e238ac23-819b-452f-9015-52922e45efd3", "external-id": "nsx-vlan-transportzone-127", "segmentation_id": 127, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap5af88b4d-69", "ovs_interfaceid": "5af88b4d-699d-4bac-b78f-f66b69bc1eea", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 862.878808] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 862.906427] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332191, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 862.955413] env[62753]: DEBUG nova.compute.utils [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 862.959604] env[62753]: DEBUG nova.compute.manager [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 862.959604] env[62753]: DEBUG nova.network.neutron [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 863.003999] env[62753]: DEBUG oslo_concurrency.lockutils [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 863.004839] env[62753]: DEBUG oslo_concurrency.lockutils [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 863.007871] env[62753]: DEBUG nova.policy [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '08b74083fa7248e7bd544fe7fecee7e8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4d0eb0b3ba504f59a85cb021da2b47b2', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 863.018762] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332192, 'name': PowerOffVM_Task, 'duration_secs': 0.154589} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.019030] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 863.019246] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 863.020026] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9c58cbb-9787-4aa9-90e6-b9f83f040560 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.027613] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 863.027905] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8dc0347e-0b5c-4420-b9f0-37a9c2afa73d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.055597] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 863.056511] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 863.056511] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Deleting the datastore file [datastore2] 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 863.056964] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-62b13d79-f4f9-4400-9bdd-e60739f80efc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.062742] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 863.062742] env[62753]: value = "task-1332194" [ 863.062742] env[62753]: _type = "Task" [ 863.062742] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.070630] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332194, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.122289] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52aa663b-e800-8093-59b8-27d943dedf8f, 'name': SearchDatastore_Task, 'duration_secs': 0.011015} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.122662] env[62753]: DEBUG oslo_concurrency.lockutils [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.122930] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 863.123198] env[62753]: DEBUG oslo_concurrency.lockutils [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 863.123393] env[62753]: DEBUG oslo_concurrency.lockutils [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 863.123532] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 863.123794] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0b7eb6d3-8a89-4f33-a132-aa7a74adbe57 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.142686] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 863.142979] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 863.143819] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5ca49fc5-bda8-4c22-b622-3bf385b02c6d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.152021] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Waiting for the task: (returnval){ [ 863.152021] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d9fa86-6f33-2769-3f10-6024d49138ee" [ 863.152021] env[62753]: _type = "Task" [ 863.152021] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.159847] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d9fa86-6f33-2769-3f10-6024d49138ee, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.352708] env[62753]: DEBUG oslo_concurrency.lockutils [req-43116918-8399-48ef-a788-401e6b80e0c4 req-ae2029de-e6b1-4891-a70e-d2e4dcbf203c service nova] Releasing lock "refresh_cache-f132230d-0fba-4293-b940-85cc7ef1cad1" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 863.384022] env[62753]: DEBUG nova.network.neutron [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 863.405877] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332191, 'name': CloneVM_Task} progress is 95%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.460089] env[62753]: DEBUG nova.compute.manager [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 863.497127] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 3bc05fdb-81be-4764-b9d9-01acd125b020 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.497303] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 046fc3dc-a187-4fda-bc66-345e1226f83d actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.497443] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.497566] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance f132230d-0fba-4293-b940-85cc7ef1cad1 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.497686] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance a6ee010c-6305-4009-80e0-92a2c58bcd7b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.497884] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 863.549218] env[62753]: DEBUG nova.network.neutron [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Updating instance_info_cache with network_info: [{"id": "bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c", "address": "fa:16:3e:0b:98:33", "network": {"id": "2823647f-d32d-4a1c-9e02-7577016a260e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1594498922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6c8a6bc946a469fa85cc3ab80d7333a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcc79f72-e1", "ovs_interfaceid": "bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 863.573261] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332194, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.112101} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.573650] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 863.574126] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 863.574219] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 863.643038] env[62753]: DEBUG nova.network.neutron [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Successfully created port: 7fb046f4-e470-4f6d-8fa3-73f3b0157020 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 863.660027] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d9fa86-6f33-2769-3f10-6024d49138ee, 'name': SearchDatastore_Task, 'duration_secs': 0.01056} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.660667] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea9374e5-d3ca-4e07-bd59-9d8e6885fb28 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.665955] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Waiting for the task: (returnval){ [ 863.665955] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52861f0d-f0ca-9867-29e5-357a8d77729b" [ 863.665955] env[62753]: _type = "Task" [ 863.665955] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 863.676707] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52861f0d-f0ca-9867-29e5-357a8d77729b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 863.905809] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332191, 'name': CloneVM_Task, 'duration_secs': 1.23006} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 863.906083] env[62753]: INFO nova.virt.vmwareapi.vmops [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Created linked-clone VM from snapshot [ 863.906872] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-757f3847-10f9-467c-8517-436a89c9d0c1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.916017] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Uploading image f136d03d-2911-4542-8963-164084c24bfe {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 863.947102] env[62753]: DEBUG oslo_vmware.rw_handles [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 863.947102] env[62753]: value = "vm-284574" [ 863.947102] env[62753]: _type = "VirtualMachine" [ 863.947102] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 863.947102] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-35623a62-d839-48aa-9700-590c9f416203 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 863.956203] env[62753]: DEBUG oslo_vmware.rw_handles [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lease: (returnval){ [ 863.956203] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529f4004-8c32-bf5a-914c-2e3941d980fd" [ 863.956203] env[62753]: _type = "HttpNfcLease" [ 863.956203] env[62753]: } obtained for exporting VM: (result){ [ 863.956203] env[62753]: value = "vm-284574" [ 863.956203] env[62753]: _type = "VirtualMachine" [ 863.956203] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 863.956203] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the lease: (returnval){ [ 863.956203] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529f4004-8c32-bf5a-914c-2e3941d980fd" [ 863.956203] env[62753]: _type = "HttpNfcLease" [ 863.956203] env[62753]: } to be ready. {{(pid=62753) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 863.961490] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 863.961490] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529f4004-8c32-bf5a-914c-2e3941d980fd" [ 863.961490] env[62753]: _type = "HttpNfcLease" [ 863.961490] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 864.001770] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 49311ef5-733a-4146-87ac-876a121a8d4d has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 864.051908] env[62753]: DEBUG oslo_concurrency.lockutils [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "refresh_cache-a6ee010c-6305-4009-80e0-92a2c58bcd7b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.053016] env[62753]: DEBUG nova.compute.manager [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Instance network_info: |[{"id": "bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c", "address": "fa:16:3e:0b:98:33", "network": {"id": "2823647f-d32d-4a1c-9e02-7577016a260e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1594498922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6c8a6bc946a469fa85cc3ab80d7333a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcc79f72-e1", "ovs_interfaceid": "bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 864.053140] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:98:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9d39252e-42ef-4252-98d3-62af5a0d109d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 864.061121] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Creating folder: Project (e6c8a6bc946a469fa85cc3ab80d7333a). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 864.061121] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9583b157-ded4-429f-a94b-7f709600a920 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.071511] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Created folder: Project (e6c8a6bc946a469fa85cc3ab80d7333a) in parent group-v284541. [ 864.071692] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Creating folder: Instances. Parent ref: group-v284575. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 864.071943] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d873bd6b-91ae-4ed3-97e5-2f904bdf0d2d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.082772] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Created folder: Instances in parent group-v284575. [ 864.083023] env[62753]: DEBUG oslo.service.loopingcall [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.083405] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 864.083611] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cd220239-e007-4ee7-8522-dcb47066685f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.101951] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 864.101951] env[62753]: value = "task-1332198" [ 864.101951] env[62753]: _type = "Task" [ 864.101951] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.109391] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332198, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.177273] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52861f0d-f0ca-9867-29e5-357a8d77729b, 'name': SearchDatastore_Task, 'duration_secs': 0.009511} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.177543] env[62753]: DEBUG oslo_concurrency.lockutils [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 864.177790] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] f132230d-0fba-4293-b940-85cc7ef1cad1/f132230d-0fba-4293-b940-85cc7ef1cad1.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 864.178091] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f4d7f014-8b03-440c-bb0d-c05a114dbd85 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.184663] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Waiting for the task: (returnval){ [ 864.184663] env[62753]: value = "task-1332199" [ 864.184663] env[62753]: _type = "Task" [ 864.184663] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.194265] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': task-1332199, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.392797] env[62753]: DEBUG nova.compute.manager [req-f6851eeb-b30f-4fdc-9a88-a99ac4cf7f16 req-7739582c-a3d0-486b-beae-9cdb5d2cc006 service nova] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Received event network-vif-plugged-bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.393449] env[62753]: DEBUG oslo_concurrency.lockutils [req-f6851eeb-b30f-4fdc-9a88-a99ac4cf7f16 req-7739582c-a3d0-486b-beae-9cdb5d2cc006 service nova] Acquiring lock "a6ee010c-6305-4009-80e0-92a2c58bcd7b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.393449] env[62753]: DEBUG oslo_concurrency.lockutils [req-f6851eeb-b30f-4fdc-9a88-a99ac4cf7f16 req-7739582c-a3d0-486b-beae-9cdb5d2cc006 service nova] Lock "a6ee010c-6305-4009-80e0-92a2c58bcd7b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 864.393569] env[62753]: DEBUG oslo_concurrency.lockutils [req-f6851eeb-b30f-4fdc-9a88-a99ac4cf7f16 req-7739582c-a3d0-486b-beae-9cdb5d2cc006 service nova] Lock "a6ee010c-6305-4009-80e0-92a2c58bcd7b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 864.393960] env[62753]: DEBUG nova.compute.manager [req-f6851eeb-b30f-4fdc-9a88-a99ac4cf7f16 req-7739582c-a3d0-486b-beae-9cdb5d2cc006 service nova] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] No waiting events found dispatching network-vif-plugged-bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 864.394644] env[62753]: WARNING nova.compute.manager [req-f6851eeb-b30f-4fdc-9a88-a99ac4cf7f16 req-7739582c-a3d0-486b-beae-9cdb5d2cc006 service nova] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Received unexpected event network-vif-plugged-bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c for instance with vm_state building and task_state spawning. [ 864.394644] env[62753]: DEBUG nova.compute.manager [req-f6851eeb-b30f-4fdc-9a88-a99ac4cf7f16 req-7739582c-a3d0-486b-beae-9cdb5d2cc006 service nova] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Received event network-changed-bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 864.394644] env[62753]: DEBUG nova.compute.manager [req-f6851eeb-b30f-4fdc-9a88-a99ac4cf7f16 req-7739582c-a3d0-486b-beae-9cdb5d2cc006 service nova] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Refreshing instance network info cache due to event network-changed-bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 864.394857] env[62753]: DEBUG oslo_concurrency.lockutils [req-f6851eeb-b30f-4fdc-9a88-a99ac4cf7f16 req-7739582c-a3d0-486b-beae-9cdb5d2cc006 service nova] Acquiring lock "refresh_cache-a6ee010c-6305-4009-80e0-92a2c58bcd7b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.394857] env[62753]: DEBUG oslo_concurrency.lockutils [req-f6851eeb-b30f-4fdc-9a88-a99ac4cf7f16 req-7739582c-a3d0-486b-beae-9cdb5d2cc006 service nova] Acquired lock "refresh_cache-a6ee010c-6305-4009-80e0-92a2c58bcd7b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.395020] env[62753]: DEBUG nova.network.neutron [req-f6851eeb-b30f-4fdc-9a88-a99ac4cf7f16 req-7739582c-a3d0-486b-beae-9cdb5d2cc006 service nova] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Refreshing network info cache for port bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 864.464783] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 864.464783] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529f4004-8c32-bf5a-914c-2e3941d980fd" [ 864.464783] env[62753]: _type = "HttpNfcLease" [ 864.464783] env[62753]: } is ready. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 864.465361] env[62753]: DEBUG oslo_vmware.rw_handles [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 864.465361] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529f4004-8c32-bf5a-914c-2e3941d980fd" [ 864.465361] env[62753]: _type = "HttpNfcLease" [ 864.465361] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 864.466060] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4d7630d-f7cd-4a8f-939a-a56f1970bffc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.471180] env[62753]: DEBUG nova.compute.manager [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 864.480748] env[62753]: DEBUG oslo_vmware.rw_handles [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c4ab87-f60a-7c06-c5c5-31154680233f/disk-0.vmdk from lease info. {{(pid=62753) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 864.481345] env[62753]: DEBUG oslo_vmware.rw_handles [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c4ab87-f60a-7c06-c5c5-31154680233f/disk-0.vmdk for reading. {{(pid=62753) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 864.540249] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance c5dae9a3-052b-4f4c-ac79-84aeaded457a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 864.550660] env[62753]: DEBUG nova.virt.hardware [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 864.550951] env[62753]: DEBUG nova.virt.hardware [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 864.551133] env[62753]: DEBUG nova.virt.hardware [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 864.551328] env[62753]: DEBUG nova.virt.hardware [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 864.551542] env[62753]: DEBUG nova.virt.hardware [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 864.552556] env[62753]: DEBUG nova.virt.hardware [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 864.552556] env[62753]: DEBUG nova.virt.hardware [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 864.552556] env[62753]: DEBUG nova.virt.hardware [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 864.552556] env[62753]: DEBUG nova.virt.hardware [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 864.552556] env[62753]: DEBUG nova.virt.hardware [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 864.552744] env[62753]: DEBUG nova.virt.hardware [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 864.553468] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8056a89a-ade1-4939-8506-3e28098a3392 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.563709] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2431046c-8ab2-455e-9adc-2989edaff4a6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.582178] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-3ef3ef9b-644e-434f-b592-aa885b64d2d3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.613472] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332198, 'name': CreateVM_Task, 'duration_secs': 0.349624} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.613472] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 864.614608] env[62753]: DEBUG oslo_concurrency.lockutils [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 864.614608] env[62753]: DEBUG oslo_concurrency.lockutils [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 864.614608] env[62753]: DEBUG oslo_concurrency.lockutils [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 864.614853] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ae6e550e-e274-4c28-be19-ec03c55c9597 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.619293] env[62753]: DEBUG nova.virt.hardware [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 864.619526] env[62753]: DEBUG nova.virt.hardware [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 864.619683] env[62753]: DEBUG nova.virt.hardware [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 864.619861] env[62753]: DEBUG nova.virt.hardware [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 864.620021] env[62753]: DEBUG nova.virt.hardware [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 864.620230] env[62753]: DEBUG nova.virt.hardware [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 864.620478] env[62753]: DEBUG nova.virt.hardware [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 864.620641] env[62753]: DEBUG nova.virt.hardware [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 864.620814] env[62753]: DEBUG nova.virt.hardware [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 864.620976] env[62753]: DEBUG nova.virt.hardware [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 864.621194] env[62753]: DEBUG nova.virt.hardware [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 864.623428] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aba6c1b7-5620-432c-b74b-ae171cf30404 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.627744] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 864.627744] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522a43d1-25aa-654f-977e-b208f63129ed" [ 864.627744] env[62753]: _type = "Task" [ 864.627744] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.638393] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d30bcdc5-eebe-4cc2-bd84-7e83b352d2f3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.651762] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522a43d1-25aa-654f-977e-b208f63129ed, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.662022] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Instance VIF info [] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 864.667649] env[62753]: DEBUG oslo.service.loopingcall [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 864.667873] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 864.668090] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0652ff71-191d-4937-b5ec-dcf2a649b495 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.684823] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 864.684823] env[62753]: value = "task-1332200" [ 864.684823] env[62753]: _type = "Task" [ 864.684823] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.696826] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332200, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.699851] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': task-1332199, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483088} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 864.700086] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] f132230d-0fba-4293-b940-85cc7ef1cad1/f132230d-0fba-4293-b940-85cc7ef1cad1.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 864.700316] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 864.700580] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fc443ca4-f8bf-4c9c-b42e-45d136ac6abe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 864.708447] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Waiting for the task: (returnval){ [ 864.708447] env[62753]: value = "task-1332201" [ 864.708447] env[62753]: _type = "Task" [ 864.708447] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 864.715629] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': task-1332201, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 864.990785] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Acquiring lock "d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 864.990785] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Lock "d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 865.045106] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 4d20e083-2959-453a-8875-47955bc02613 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 865.142566] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522a43d1-25aa-654f-977e-b208f63129ed, 'name': SearchDatastore_Task, 'duration_secs': 0.030569} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.143266] env[62753]: DEBUG oslo_concurrency.lockutils [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.143881] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.143881] env[62753]: DEBUG oslo_concurrency.lockutils [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.144031] env[62753]: DEBUG oslo_concurrency.lockutils [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.144519] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 865.145300] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cdb2ac73-ceda-467f-880f-79eeb31275ff {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.154609] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 865.154795] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 865.155565] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8021ad24-1a67-48a2-8f46-63f20149b057 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.163069] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 865.163069] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522ef525-437b-d63f-9fdc-54daaf0bf29e" [ 865.163069] env[62753]: _type = "Task" [ 865.163069] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.170641] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522ef525-437b-d63f-9fdc-54daaf0bf29e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.196178] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332200, 'name': CreateVM_Task, 'duration_secs': 0.265822} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.196438] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 865.196966] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.197230] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.197581] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 865.197804] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dd6477ce-975b-4874-a64c-d17489e4bd06 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.202854] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 865.202854] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522f2c0a-40fc-8b85-d03a-b273624cf03e" [ 865.202854] env[62753]: _type = "Task" [ 865.202854] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.211215] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522f2c0a-40fc-8b85-d03a-b273624cf03e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.220266] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': task-1332201, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.095957} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.220266] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 865.220266] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-afa0dda5-179a-49ca-b9af-c82d00586c3f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.242977] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Reconfiguring VM instance instance-0000003c to attach disk [datastore1] f132230d-0fba-4293-b940-85cc7ef1cad1/f132230d-0fba-4293-b940-85cc7ef1cad1.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 865.243366] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-db160b7d-f4a2-4e06-a00c-68896ae3347f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.267267] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Waiting for the task: (returnval){ [ 865.267267] env[62753]: value = "task-1332202" [ 865.267267] env[62753]: _type = "Task" [ 865.267267] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.277310] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': task-1332202, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.318409] env[62753]: DEBUG nova.network.neutron [req-f6851eeb-b30f-4fdc-9a88-a99ac4cf7f16 req-7739582c-a3d0-486b-beae-9cdb5d2cc006 service nova] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Updated VIF entry in instance network info cache for port bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 865.318801] env[62753]: DEBUG nova.network.neutron [req-f6851eeb-b30f-4fdc-9a88-a99ac4cf7f16 req-7739582c-a3d0-486b-beae-9cdb5d2cc006 service nova] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Updating instance_info_cache with network_info: [{"id": "bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c", "address": "fa:16:3e:0b:98:33", "network": {"id": "2823647f-d32d-4a1c-9e02-7577016a260e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1594498922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6c8a6bc946a469fa85cc3ab80d7333a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapbcc79f72-e1", "ovs_interfaceid": "bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 865.324335] env[62753]: DEBUG nova.network.neutron [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Successfully updated port: 7fb046f4-e470-4f6d-8fa3-73f3b0157020 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 865.549988] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 50e1c1a1-7b8f-49cd-932d-03c920209634 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 865.672946] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522ef525-437b-d63f-9fdc-54daaf0bf29e, 'name': SearchDatastore_Task, 'duration_secs': 0.009512} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.673894] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9d5b1498-138b-4700-a1f9-33fb71aa289f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.679414] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 865.679414] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522b2994-1923-7ceb-d335-33c4aefa2b78" [ 865.679414] env[62753]: _type = "Task" [ 865.679414] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.687624] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522b2994-1923-7ceb-d335-33c4aefa2b78, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.712362] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522f2c0a-40fc-8b85-d03a-b273624cf03e, 'name': SearchDatastore_Task, 'duration_secs': 0.009332} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.712709] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.713067] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 865.713637] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.776918] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': task-1332202, 'name': ReconfigVM_Task, 'duration_secs': 0.465096} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 865.777306] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Reconfigured VM instance instance-0000003c to attach disk [datastore1] f132230d-0fba-4293-b940-85cc7ef1cad1/f132230d-0fba-4293-b940-85cc7ef1cad1.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 865.778056] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-0f23e7ce-125c-4cc0-8205-3e00988d41be {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 865.784021] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Waiting for the task: (returnval){ [ 865.784021] env[62753]: value = "task-1332203" [ 865.784021] env[62753]: _type = "Task" [ 865.784021] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 865.792734] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': task-1332203, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 865.824481] env[62753]: DEBUG oslo_concurrency.lockutils [req-f6851eeb-b30f-4fdc-9a88-a99ac4cf7f16 req-7739582c-a3d0-486b-beae-9cdb5d2cc006 service nova] Releasing lock "refresh_cache-a6ee010c-6305-4009-80e0-92a2c58bcd7b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 865.826508] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 865.826508] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquired lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 865.826620] env[62753]: DEBUG nova.network.neutron [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 866.052643] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 33c1de85-aade-4c69-8fdb-7672d8c53300 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 866.190147] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522b2994-1923-7ceb-d335-33c4aefa2b78, 'name': SearchDatastore_Task, 'duration_secs': 0.012739} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.190468] env[62753]: DEBUG oslo_concurrency.lockutils [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 866.190768] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] a6ee010c-6305-4009-80e0-92a2c58bcd7b/a6ee010c-6305-4009-80e0-92a2c58bcd7b.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 866.191092] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 866.191295] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 866.191548] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d44661fc-504d-44cb-a9b5-1569dfe9cab5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.193653] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-410b6488-3807-43dc-9a1f-3f9006859003 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.200382] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 866.200382] env[62753]: value = "task-1332204" [ 866.200382] env[62753]: _type = "Task" [ 866.200382] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.204443] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 866.204647] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 866.205727] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2c6485d0-4820-4895-b87f-ec4595599eb3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.210997] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332204, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.213910] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 866.213910] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522904cb-3369-4448-8cb1-7ba816595c0b" [ 866.213910] env[62753]: _type = "Task" [ 866.213910] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.221217] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522904cb-3369-4448-8cb1-7ba816595c0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.294654] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': task-1332203, 'name': Rename_Task, 'duration_secs': 0.148201} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.295143] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 866.295412] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4ad37137-56b8-43c7-8ab5-e76e1963850a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.301240] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Waiting for the task: (returnval){ [ 866.301240] env[62753]: value = "task-1332205" [ 866.301240] env[62753]: _type = "Task" [ 866.301240] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.308849] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': task-1332205, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.361292] env[62753]: DEBUG nova.network.neutron [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 866.430910] env[62753]: DEBUG nova.compute.manager [req-e3526dba-ff14-47e1-a21f-acfd7088d96a req-fac91c52-f984-49d2-845b-65b2f2ee97ee service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Received event network-vif-plugged-7fb046f4-e470-4f6d-8fa3-73f3b0157020 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.431155] env[62753]: DEBUG oslo_concurrency.lockutils [req-e3526dba-ff14-47e1-a21f-acfd7088d96a req-fac91c52-f984-49d2-845b-65b2f2ee97ee service nova] Acquiring lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 866.431510] env[62753]: DEBUG oslo_concurrency.lockutils [req-e3526dba-ff14-47e1-a21f-acfd7088d96a req-fac91c52-f984-49d2-845b-65b2f2ee97ee service nova] Lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 866.431629] env[62753]: DEBUG oslo_concurrency.lockutils [req-e3526dba-ff14-47e1-a21f-acfd7088d96a req-fac91c52-f984-49d2-845b-65b2f2ee97ee service nova] Lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 866.431815] env[62753]: DEBUG nova.compute.manager [req-e3526dba-ff14-47e1-a21f-acfd7088d96a req-fac91c52-f984-49d2-845b-65b2f2ee97ee service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] No waiting events found dispatching network-vif-plugged-7fb046f4-e470-4f6d-8fa3-73f3b0157020 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 866.432018] env[62753]: WARNING nova.compute.manager [req-e3526dba-ff14-47e1-a21f-acfd7088d96a req-fac91c52-f984-49d2-845b-65b2f2ee97ee service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Received unexpected event network-vif-plugged-7fb046f4-e470-4f6d-8fa3-73f3b0157020 for instance with vm_state building and task_state spawning. [ 866.432193] env[62753]: DEBUG nova.compute.manager [req-e3526dba-ff14-47e1-a21f-acfd7088d96a req-fac91c52-f984-49d2-845b-65b2f2ee97ee service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Received event network-changed-7fb046f4-e470-4f6d-8fa3-73f3b0157020 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 866.432354] env[62753]: DEBUG nova.compute.manager [req-e3526dba-ff14-47e1-a21f-acfd7088d96a req-fac91c52-f984-49d2-845b-65b2f2ee97ee service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Refreshing instance network info cache due to event network-changed-7fb046f4-e470-4f6d-8fa3-73f3b0157020. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 866.432533] env[62753]: DEBUG oslo_concurrency.lockutils [req-e3526dba-ff14-47e1-a21f-acfd7088d96a req-fac91c52-f984-49d2-845b-65b2f2ee97ee service nova] Acquiring lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 866.545444] env[62753]: DEBUG nova.network.neutron [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Updating instance_info_cache with network_info: [{"id": "7fb046f4-e470-4f6d-8fa3-73f3b0157020", "address": "fa:16:3e:6f:26:2c", "network": {"id": "fc646f7f-70ba-4671-8476-3b0ece1e9041", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-853174775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d0eb0b3ba504f59a85cb021da2b47b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fb046f4-e4", "ovs_interfaceid": "7fb046f4-e470-4f6d-8fa3-73f3b0157020", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 866.560357] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 57eef548-a0f1-4f0b-a026-885a10d005c2 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 866.711138] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332204, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483363} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.711429] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] a6ee010c-6305-4009-80e0-92a2c58bcd7b/a6ee010c-6305-4009-80e0-92a2c58bcd7b.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 866.711662] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 866.711913] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fcbc6ddc-57fe-4027-a2f2-ac8ee219701d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.718524] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 866.718524] env[62753]: value = "task-1332206" [ 866.718524] env[62753]: _type = "Task" [ 866.718524] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.725426] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522904cb-3369-4448-8cb1-7ba816595c0b, 'name': SearchDatastore_Task, 'duration_secs': 0.008266} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 866.726752] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c47fde45-8021-4528-bc9a-2deb69944f9b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 866.732337] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332206, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.735594] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 866.735594] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529b1f6e-0e9e-0ccf-700f-926ea6090c02" [ 866.735594] env[62753]: _type = "Task" [ 866.735594] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 866.744204] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529b1f6e-0e9e-0ccf-700f-926ea6090c02, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 866.812376] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': task-1332205, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.048458] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Releasing lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.049167] env[62753]: DEBUG nova.compute.manager [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Instance network_info: |[{"id": "7fb046f4-e470-4f6d-8fa3-73f3b0157020", "address": "fa:16:3e:6f:26:2c", "network": {"id": "fc646f7f-70ba-4671-8476-3b0ece1e9041", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-853174775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d0eb0b3ba504f59a85cb021da2b47b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fb046f4-e4", "ovs_interfaceid": "7fb046f4-e470-4f6d-8fa3-73f3b0157020", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 867.049263] env[62753]: DEBUG oslo_concurrency.lockutils [req-e3526dba-ff14-47e1-a21f-acfd7088d96a req-fac91c52-f984-49d2-845b-65b2f2ee97ee service nova] Acquired lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.049413] env[62753]: DEBUG nova.network.neutron [req-e3526dba-ff14-47e1-a21f-acfd7088d96a req-fac91c52-f984-49d2-845b-65b2f2ee97ee service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Refreshing network info cache for port 7fb046f4-e470-4f6d-8fa3-73f3b0157020 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 867.050691] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:26:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '255460d5-71d4-4bfd-87f1-acc10085db7f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7fb046f4-e470-4f6d-8fa3-73f3b0157020', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 867.058584] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Creating folder: Project (4d0eb0b3ba504f59a85cb021da2b47b2). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 867.061613] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-ffda8514-f330-4960-bc29-c0cf60edd786 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.064018] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance f03739f4-921f-46e5-b0e7-e9a99d96d621 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 867.075118] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Created folder: Project (4d0eb0b3ba504f59a85cb021da2b47b2) in parent group-v284541. [ 867.075341] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Creating folder: Instances. Parent ref: group-v284579. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 867.075585] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9c9bc163-d06c-4508-adaa-dc159636e951 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.085592] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Created folder: Instances in parent group-v284579. [ 867.085931] env[62753]: DEBUG oslo.service.loopingcall [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 867.086047] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 867.086266] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dd3483e9-9110-4024-87e1-91d0215f0bc5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.108471] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 867.108471] env[62753]: value = "task-1332209" [ 867.108471] env[62753]: _type = "Task" [ 867.108471] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.117082] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332209, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.230417] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332206, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.063352} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.230690] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 867.231500] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04a8748f-b7d7-4602-85a4-17862e1d01ff {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.259507] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] a6ee010c-6305-4009-80e0-92a2c58bcd7b/a6ee010c-6305-4009-80e0-92a2c58bcd7b.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 867.260593] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-95002fc9-d7e2-4af3-a04c-6d1e3f742404 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.280243] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529b1f6e-0e9e-0ccf-700f-926ea6090c02, 'name': SearchDatastore_Task, 'duration_secs': 0.008228} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.280844] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.281155] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3/4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 867.281430] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d755cc7d-22a8-473d-ad78-4b0b8516def8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.286192] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 867.286192] env[62753]: value = "task-1332210" [ 867.286192] env[62753]: _type = "Task" [ 867.286192] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.292034] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 867.292034] env[62753]: value = "task-1332211" [ 867.292034] env[62753]: _type = "Task" [ 867.292034] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.297498] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332210, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.299452] env[62753]: DEBUG nova.network.neutron [req-e3526dba-ff14-47e1-a21f-acfd7088d96a req-fac91c52-f984-49d2-845b-65b2f2ee97ee service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Updated VIF entry in instance network info cache for port 7fb046f4-e470-4f6d-8fa3-73f3b0157020. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 867.299808] env[62753]: DEBUG nova.network.neutron [req-e3526dba-ff14-47e1-a21f-acfd7088d96a req-fac91c52-f984-49d2-845b-65b2f2ee97ee service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Updating instance_info_cache with network_info: [{"id": "7fb046f4-e470-4f6d-8fa3-73f3b0157020", "address": "fa:16:3e:6f:26:2c", "network": {"id": "fc646f7f-70ba-4671-8476-3b0ece1e9041", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-853174775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d0eb0b3ba504f59a85cb021da2b47b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fb046f4-e4", "ovs_interfaceid": "7fb046f4-e470-4f6d-8fa3-73f3b0157020", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 867.303887] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332211, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.312854] env[62753]: DEBUG oslo_vmware.api [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': task-1332205, 'name': PowerOnVM_Task, 'duration_secs': 0.615927} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.313823] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 867.314053] env[62753]: INFO nova.compute.manager [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Took 7.66 seconds to spawn the instance on the hypervisor. [ 867.314252] env[62753]: DEBUG nova.compute.manager [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 867.315104] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f207a2a-6443-4baf-ae4b-d9c905d96827 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.568011] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 867.620097] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332209, 'name': CreateVM_Task, 'duration_secs': 0.393722} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.620316] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 867.621171] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 867.621347] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 867.621715] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 867.621986] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a81eaf9-fc7f-4e28-875f-eec8e2a6715d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.630186] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 867.630186] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52170126-5440-68c5-b442-8111b5859c29" [ 867.630186] env[62753]: _type = "Task" [ 867.630186] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.638670] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52170126-5440-68c5-b442-8111b5859c29, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.799476] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332210, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.805118] env[62753]: DEBUG oslo_concurrency.lockutils [req-e3526dba-ff14-47e1-a21f-acfd7088d96a req-fac91c52-f984-49d2-845b-65b2f2ee97ee service nova] Releasing lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 867.805556] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332211, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.478963} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 867.805820] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3/4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 867.806066] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 867.806358] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-82166074-10b0-47bc-b0e5-8ed7d18f486a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 867.812670] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 867.812670] env[62753]: value = "task-1332212" [ 867.812670] env[62753]: _type = "Task" [ 867.812670] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 867.820665] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332212, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 867.833035] env[62753]: INFO nova.compute.manager [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Took 30.83 seconds to build instance. [ 868.071772] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 6b9a08b0-3cb3-474a-8193-953e74252a53 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 868.141190] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52170126-5440-68c5-b442-8111b5859c29, 'name': SearchDatastore_Task, 'duration_secs': 0.060792} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.141606] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 868.141907] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 868.142597] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 868.142597] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 868.142803] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 868.142969] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-33a9920f-e83c-4477-a1ca-788cef945f51 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.152610] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 868.153438] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 868.154292] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2047c5ea-700f-4e4e-8296-8f40d069505e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.160331] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 868.160331] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a6cbf1-27e2-f21e-724d-389a900ad779" [ 868.160331] env[62753]: _type = "Task" [ 868.160331] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.169492] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a6cbf1-27e2-f21e-724d-389a900ad779, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.298098] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332210, 'name': ReconfigVM_Task, 'duration_secs': 0.555279} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.298417] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Reconfigured VM instance instance-0000003d to attach disk [datastore1] a6ee010c-6305-4009-80e0-92a2c58bcd7b/a6ee010c-6305-4009-80e0-92a2c58bcd7b.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 868.299068] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-683a1c1c-12ed-471c-92d7-dfe4696179db {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.305387] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 868.305387] env[62753]: value = "task-1332213" [ 868.305387] env[62753]: _type = "Task" [ 868.305387] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.314215] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332213, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.323667] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332212, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068544} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.323971] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 868.324817] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da243535-da5b-4bf7-b048-1b1718ea0e46 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.338055] env[62753]: DEBUG oslo_concurrency.lockutils [None req-712031fb-1121-4446-b612-868bd14ec311 tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Lock "f132230d-0fba-4293-b940-85cc7ef1cad1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 83.311s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 868.347090] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Reconfiguring VM instance instance-0000003b to attach disk [datastore1] 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3/4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 868.348917] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f7ecd466-d781-4a60-8cbf-bb621dff4f4e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.364861] env[62753]: DEBUG nova.compute.manager [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 868.372644] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 868.372644] env[62753]: value = "task-1332214" [ 868.372644] env[62753]: _type = "Task" [ 868.372644] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.382786] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332214, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.575236] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance c73b1ae0-dc98-47f7-babf-e96169384785 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 868.672724] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a6cbf1-27e2-f21e-724d-389a900ad779, 'name': SearchDatastore_Task, 'duration_secs': 0.015821} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.672724] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef20c2b8-f0b7-4e12-a9c6-b45dc41b977a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.677813] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 868.677813] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5276ccfe-069a-7421-5c1e-a3ee73873d14" [ 868.677813] env[62753]: _type = "Task" [ 868.677813] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.686695] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5276ccfe-069a-7421-5c1e-a3ee73873d14, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.817041] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332213, 'name': Rename_Task, 'duration_secs': 0.206238} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.817041] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 868.817448] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-270353a1-560d-46da-aeea-4809f4b61f1e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.824312] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 868.824312] env[62753]: value = "task-1332215" [ 868.824312] env[62753]: _type = "Task" [ 868.824312] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.832563] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332215, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 868.885223] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332214, 'name': ReconfigVM_Task, 'duration_secs': 0.383145} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 868.885518] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Reconfigured VM instance instance-0000003b to attach disk [datastore1] 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3/4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 868.886185] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-db47f580-3ff9-4d64-a58f-058a668b2e20 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 868.888737] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 868.893660] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 868.893660] env[62753]: value = "task-1332216" [ 868.893660] env[62753]: _type = "Task" [ 868.893660] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 868.903339] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332216, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.079571] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 77cdd901-cc96-4a6f-a696-a65f54a96b1a has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 869.079837] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 869.080029] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 869.192314] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5276ccfe-069a-7421-5c1e-a3ee73873d14, 'name': SearchDatastore_Task, 'duration_secs': 0.009027} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.192793] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 869.192904] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5/4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 869.193313] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5b899cf5-e4a1-4fe6-95fa-14ddd4d7091e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.203029] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 869.203029] env[62753]: value = "task-1332217" [ 869.203029] env[62753]: _type = "Task" [ 869.203029] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.213682] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332217, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.335812] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b468f0-9470-4c68-975c-0de01d2b85a1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.339551] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332215, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.344021] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a51e97f2-46ea-4133-b0d3-27d3897036bd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.375405] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb63b8e6-6de4-4649-b4c6-60052a4370ec {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.383383] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf628c17-518a-4549-9b3d-4713114b5719 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.397283] env[62753]: DEBUG nova.compute.provider_tree [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 869.409067] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332216, 'name': Rename_Task, 'duration_secs': 0.197587} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.410204] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 869.410522] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-043639f4-8b91-43e8-9039-47ddf860e4e6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.419962] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 869.419962] env[62753]: value = "task-1332218" [ 869.419962] env[62753]: _type = "Task" [ 869.419962] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 869.429565] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332218, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.717950] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332217, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 869.835435] env[62753]: DEBUG oslo_vmware.api [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332215, 'name': PowerOnVM_Task, 'duration_secs': 0.791849} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 869.835435] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 869.835656] env[62753]: INFO nova.compute.manager [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Took 7.86 seconds to spawn the instance on the hypervisor. [ 869.835838] env[62753]: DEBUG nova.compute.manager [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 869.836636] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f60271b-3302-440b-a076-16a940ee06e8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 869.904719] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 869.931289] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332218, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.006660] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Acquiring lock "f132230d-0fba-4293-b940-85cc7ef1cad1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.006947] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Lock "f132230d-0fba-4293-b940-85cc7ef1cad1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.007195] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Acquiring lock "f132230d-0fba-4293-b940-85cc7ef1cad1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 870.007396] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Lock "f132230d-0fba-4293-b940-85cc7ef1cad1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.007573] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Lock "f132230d-0fba-4293-b940-85cc7ef1cad1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.010098] env[62753]: INFO nova.compute.manager [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Terminating instance [ 870.012007] env[62753]: DEBUG nova.compute.manager [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 870.012274] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 870.013226] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2886b36b-42ad-4a7c-965b-8db78def0aff {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.020947] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 870.021233] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1dd6e219-1e64-4019-b569-7f0f6564637b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.028083] env[62753]: DEBUG oslo_vmware.api [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Waiting for the task: (returnval){ [ 870.028083] env[62753]: value = "task-1332219" [ 870.028083] env[62753]: _type = "Task" [ 870.028083] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.036356] env[62753]: DEBUG oslo_vmware.api [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': task-1332219, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.212709] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332217, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.53215} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.212999] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5/4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 870.213237] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 870.213497] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0c092e2f-bfa7-4ad1-a807-c9915a87d9ad {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.219758] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 870.219758] env[62753]: value = "task-1332220" [ 870.219758] env[62753]: _type = "Task" [ 870.219758] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.227550] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332220, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.353424] env[62753]: INFO nova.compute.manager [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Took 30.75 seconds to build instance. [ 870.409640] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 870.409864] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 7.962s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.410163] env[62753]: DEBUG oslo_concurrency.lockutils [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.634s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 870.411689] env[62753]: INFO nova.compute.claims [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 870.429863] env[62753]: DEBUG oslo_vmware.api [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332218, 'name': PowerOnVM_Task, 'duration_secs': 0.589812} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.429946] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 870.430160] env[62753]: DEBUG nova.compute.manager [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 870.431434] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aac389e-a4a8-4bd0-b32d-cb3125a7cf04 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.537538] env[62753]: DEBUG oslo_vmware.api [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': task-1332219, 'name': PowerOffVM_Task, 'duration_secs': 0.340959} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.537817] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 870.537991] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 870.538292] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-66b98eac-322f-40b3-a622-0279ce04ae47 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.618304] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 870.618548] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 870.618734] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Deleting the datastore file [datastore1] f132230d-0fba-4293-b940-85cc7ef1cad1 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 870.619097] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-99ee7893-31e2-40ca-b2aa-26722f2c69cc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.624995] env[62753]: DEBUG oslo_vmware.api [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Waiting for the task: (returnval){ [ 870.624995] env[62753]: value = "task-1332222" [ 870.624995] env[62753]: _type = "Task" [ 870.624995] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.632605] env[62753]: DEBUG oslo_vmware.api [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': task-1332222, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.731617] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332220, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.11917} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 870.731920] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 870.732703] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06e76c77-bbb3-41ea-a1da-968aa5b21323 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.754319] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5/4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 870.754618] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-44f53481-bf3a-46fe-8d55-784503aa32ba {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 870.774385] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 870.774385] env[62753]: value = "task-1332223" [ 870.774385] env[62753]: _type = "Task" [ 870.774385] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 870.781979] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332223, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 870.855820] env[62753]: DEBUG oslo_concurrency.lockutils [None req-35cb0013-412b-4611-9910-bab0684d2125 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "a6ee010c-6305-4009-80e0-92a2c58bcd7b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 85.272s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 870.949038] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.135174] env[62753]: DEBUG oslo_vmware.api [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Task: {'id': task-1332222, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.151446} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.135452] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 871.135649] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 871.135964] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 871.136218] env[62753]: INFO nova.compute.manager [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Took 1.12 seconds to destroy the instance on the hypervisor. [ 871.136476] env[62753]: DEBUG oslo.service.loopingcall [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 871.136677] env[62753]: DEBUG nova.compute.manager [-] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 871.136776] env[62753]: DEBUG nova.network.neutron [-] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 871.284702] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332223, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.358567] env[62753]: DEBUG nova.compute.manager [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 871.540290] env[62753]: DEBUG nova.compute.manager [req-f47cd3bd-01ae-41b7-9f2f-8405edb18192 req-9a9954a3-5730-4378-9785-3c7dd0a14b23 service nova] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Received event network-vif-deleted-5af88b4d-699d-4bac-b78f-f66b69bc1eea {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 871.540495] env[62753]: INFO nova.compute.manager [req-f47cd3bd-01ae-41b7-9f2f-8405edb18192 req-9a9954a3-5730-4378-9785-3c7dd0a14b23 service nova] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Neutron deleted interface 5af88b4d-699d-4bac-b78f-f66b69bc1eea; detaching it from the instance and deleting it from the info cache [ 871.540738] env[62753]: DEBUG nova.network.neutron [req-f47cd3bd-01ae-41b7-9f2f-8405edb18192 req-9a9954a3-5730-4378-9785-3c7dd0a14b23 service nova] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.664916] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b76b91a-bcd5-4fcc-824b-2a8e9f12bbca {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.673054] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6cbbe03-fbb8-49c9-beb1-41d2226ebc2b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.706015] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58219009-e644-4716-8208-ea9e35e3f0bc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.712180] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d0162b5-ab4f-4a36-96e8-2ac42e196b0d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.725981] env[62753]: DEBUG nova.compute.provider_tree [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 871.784736] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332223, 'name': ReconfigVM_Task, 'duration_secs': 0.724178} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 871.785076] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5/4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 871.785731] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-aed3f858-4b38-4f88-8fd7-4a1bada2178c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 871.792186] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 871.792186] env[62753]: value = "task-1332224" [ 871.792186] env[62753]: _type = "Task" [ 871.792186] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 871.800134] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332224, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 871.877431] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.906028] env[62753]: DEBUG nova.network.neutron [-] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 871.960458] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquiring lock "4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.960798] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Lock "4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.961032] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquiring lock "4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 871.961272] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Lock "4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 871.961484] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Lock "4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 871.963820] env[62753]: INFO nova.compute.manager [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Terminating instance [ 871.965579] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquiring lock "refresh_cache-4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 871.965744] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquired lock "refresh_cache-4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 871.965917] env[62753]: DEBUG nova.network.neutron [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 872.043094] env[62753]: INFO nova.compute.manager [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Rebuilding instance [ 872.045760] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2d19bdc2-28eb-4191-a2d8-674bafd5f15c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.055897] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c07fb890-af21-400f-b844-566c36480661 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.087825] env[62753]: DEBUG nova.compute.manager [req-f47cd3bd-01ae-41b7-9f2f-8405edb18192 req-9a9954a3-5730-4378-9785-3c7dd0a14b23 service nova] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Detach interface failed, port_id=5af88b4d-699d-4bac-b78f-f66b69bc1eea, reason: Instance f132230d-0fba-4293-b940-85cc7ef1cad1 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 872.094055] env[62753]: DEBUG nova.compute.manager [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 872.094955] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f3a9061-a7f4-44a6-8da9-1ea233a3701c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.230030] env[62753]: DEBUG nova.scheduler.client.report [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 872.302994] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332224, 'name': Rename_Task, 'duration_secs': 0.331604} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 872.303334] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 872.303613] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f6e0bcec-7f1b-4d3b-9e36-466fae163b61 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.310471] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 872.310471] env[62753]: value = "task-1332225" [ 872.310471] env[62753]: _type = "Task" [ 872.310471] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.318653] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332225, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.392163] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 872.392404] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Starting heal instance info cache {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 872.392488] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Rebuilding the list of instances to heal {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 872.408973] env[62753]: INFO nova.compute.manager [-] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Took 1.27 seconds to deallocate network for instance. [ 872.484414] env[62753]: DEBUG nova.network.neutron [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 872.502941] env[62753]: DEBUG oslo_vmware.rw_handles [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c4ab87-f60a-7c06-c5c5-31154680233f/disk-0.vmdk. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 872.504013] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d350698-1699-459a-8485-cbd027f8f3da {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.510696] env[62753]: DEBUG oslo_vmware.rw_handles [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c4ab87-f60a-7c06-c5c5-31154680233f/disk-0.vmdk is in state: ready. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 872.510891] env[62753]: ERROR oslo_vmware.rw_handles [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c4ab87-f60a-7c06-c5c5-31154680233f/disk-0.vmdk due to incomplete transfer. [ 872.511152] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-ca012d4c-848c-4090-9524-7b3b92605e6e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.517996] env[62753]: DEBUG oslo_vmware.rw_handles [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52c4ab87-f60a-7c06-c5c5-31154680233f/disk-0.vmdk. {{(pid=62753) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 872.518225] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Uploaded image f136d03d-2911-4542-8963-164084c24bfe to the Glance image server {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 872.520305] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Destroying the VM {{(pid=62753) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 872.520565] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-d2358c30-4713-4dfe-b13d-490322ea066b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.526414] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 872.526414] env[62753]: value = "task-1332226" [ 872.526414] env[62753]: _type = "Task" [ 872.526414] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.534723] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332226, 'name': Destroy_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.543246] env[62753]: DEBUG nova.network.neutron [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 872.606102] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 872.606419] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c6eeebeb-38f1-48ba-b458-42d1ed92e925 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 872.613430] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 872.613430] env[62753]: value = "task-1332227" [ 872.613430] env[62753]: _type = "Task" [ 872.613430] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 872.622373] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332227, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.734775] env[62753]: DEBUG oslo_concurrency.lockutils [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.324s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 872.735508] env[62753]: DEBUG nova.compute.manager [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 872.739827] env[62753]: DEBUG oslo_concurrency.lockutils [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.066s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 872.742040] env[62753]: INFO nova.compute.claims [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 872.822954] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332225, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 872.896643] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Skipping network cache update for instance because it is being deleted. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 872.896814] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Skipping network cache update for instance because it is being deleted. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 872.896954] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 872.897094] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 872.916428] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 872.929806] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 872.929946] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquired lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 872.930129] env[62753]: DEBUG nova.network.neutron [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Forcefully refreshing network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 872.930288] env[62753]: DEBUG nova.objects.instance [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lazy-loading 'info_cache' on Instance uuid 3bc05fdb-81be-4764-b9d9-01acd125b020 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 873.036685] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332226, 'name': Destroy_Task, 'duration_secs': 0.424613} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.036977] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Destroyed the VM [ 873.037408] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Deleting Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 873.037710] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-720eb651-c659-41ae-90eb-f1a7f0ff6690 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.044894] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 873.044894] env[62753]: value = "task-1332228" [ 873.044894] env[62753]: _type = "Task" [ 873.044894] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.048276] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Releasing lock "refresh_cache-4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 873.048673] env[62753]: DEBUG nova.compute.manager [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 873.048869] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 873.049635] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-772f939b-4f50-4a39-a4b5-fd2943357c31 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.056899] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332228, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.058977] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 873.059223] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-11c64540-1a25-49bd-a7e9-ccd191fd9ebe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.064169] env[62753]: DEBUG oslo_vmware.api [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 873.064169] env[62753]: value = "task-1332229" [ 873.064169] env[62753]: _type = "Task" [ 873.064169] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.072276] env[62753]: DEBUG oslo_vmware.api [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332229, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.123357] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332227, 'name': PowerOffVM_Task, 'duration_secs': 0.1771} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.123694] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 873.124054] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 873.124910] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a0c87782-723c-461d-9eb1-9562d0b5a724 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.132087] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 873.132331] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5f3e0b04-8f3b-4148-99cd-055a6e59e188 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.190263] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 873.190756] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 873.190756] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Deleting the datastore file [datastore1] a6ee010c-6305-4009-80e0-92a2c58bcd7b {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.191128] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-37da90c1-1a7a-4eaf-9074-80acc4dd620a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.197595] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 873.197595] env[62753]: value = "task-1332231" [ 873.197595] env[62753]: _type = "Task" [ 873.197595] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.205699] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332231, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.246456] env[62753]: DEBUG nova.compute.utils [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 873.250235] env[62753]: DEBUG nova.compute.manager [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 873.250430] env[62753]: DEBUG nova.network.neutron [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 873.299938] env[62753]: DEBUG nova.policy [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4603d0d0a2a14e419f6c34d770e562e5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b01e80a17e843a98ee662157230ae17', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 873.322859] env[62753]: DEBUG oslo_vmware.api [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332225, 'name': PowerOnVM_Task, 'duration_secs': 0.628203} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.323274] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 873.323569] env[62753]: INFO nova.compute.manager [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Took 8.85 seconds to spawn the instance on the hypervisor. [ 873.323857] env[62753]: DEBUG nova.compute.manager [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 873.325310] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dffb4dc9-e604-4da0-9ab2-8147c258d3c5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.503418] env[62753]: DEBUG oslo_concurrency.lockutils [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "046fc3dc-a187-4fda-bc66-345e1226f83d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.503706] env[62753]: DEBUG oslo_concurrency.lockutils [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "046fc3dc-a187-4fda-bc66-345e1226f83d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.503923] env[62753]: DEBUG oslo_concurrency.lockutils [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "046fc3dc-a187-4fda-bc66-345e1226f83d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 873.504130] env[62753]: DEBUG oslo_concurrency.lockutils [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "046fc3dc-a187-4fda-bc66-345e1226f83d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 873.504307] env[62753]: DEBUG oslo_concurrency.lockutils [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "046fc3dc-a187-4fda-bc66-345e1226f83d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 873.507355] env[62753]: INFO nova.compute.manager [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Terminating instance [ 873.510320] env[62753]: DEBUG nova.compute.manager [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 873.510559] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 873.511555] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d9731c5-1d8a-42fe-a902-189c1f1a3ef1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.519564] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 873.519809] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-adfbd153-9202-4487-a8ec-9292532cec0e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.555480] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332228, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.572760] env[62753]: DEBUG oslo_vmware.api [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332229, 'name': PowerOffVM_Task, 'duration_secs': 0.269809} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.573419] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 873.573605] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 873.573889] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b3b32228-08d7-48f6-9762-d7b43c4006c1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.576364] env[62753]: DEBUG nova.network.neutron [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Successfully created port: 2b4dd6f0-5b5b-47ee-b7cb-ecafe1190758 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 873.600622] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 873.600890] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 873.601091] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Deleting the datastore file [datastore1] 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 873.601363] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dbfdb07d-3d02-4e6d-bfa2-0c14b003640e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 873.608058] env[62753]: DEBUG oslo_vmware.api [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for the task: (returnval){ [ 873.608058] env[62753]: value = "task-1332234" [ 873.608058] env[62753]: _type = "Task" [ 873.608058] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 873.615877] env[62753]: DEBUG oslo_vmware.api [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332234, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 873.713789] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332231, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15307} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 873.714130] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 873.714332] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 873.714511] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 873.755528] env[62753]: DEBUG nova.compute.manager [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 873.848177] env[62753]: INFO nova.compute.manager [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Took 31.68 seconds to build instance. [ 873.859010] env[62753]: DEBUG nova.network.neutron [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Successfully created port: 51f3d79a-315b-418b-ac51-a3b5f5c328f3 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 874.035226] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "b5674964-1928-4ecb-b1fd-8f60a94b4270" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 874.035587] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "b5674964-1928-4ecb-b1fd-8f60a94b4270" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 874.041798] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd28f75a-2cb8-46e9-83b2-c5901e2ff135 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.054270] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d5482b5-7a06-4e98-ba11-1a977c07176b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.066015] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332228, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.092352] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3fff85d-7d07-4dd8-beb4-0ea9fa3fb63b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.101019] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13ab1b54-f022-46db-a3f8-eceaf5286804 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.117976] env[62753]: DEBUG nova.compute.provider_tree [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 874.130027] env[62753]: DEBUG oslo_vmware.api [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Task: {'id': task-1332234, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.103025} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.130027] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 874.130027] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 874.130027] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 874.130027] env[62753]: INFO nova.compute.manager [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Took 1.08 seconds to destroy the instance on the hypervisor. [ 874.130317] env[62753]: DEBUG oslo.service.loopingcall [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.130706] env[62753]: DEBUG nova.compute.manager [-] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 874.130849] env[62753]: DEBUG nova.network.neutron [-] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 874.146315] env[62753]: DEBUG nova.network.neutron [-] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 874.200458] env[62753]: DEBUG nova.network.neutron [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Successfully created port: f07bc1fd-c64f-4525-b198-0d7cc81d6927 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 874.350605] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aee0f8e7-4590-4d2e-a6f8-b14f1d17f57a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 84.315s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 874.556058] env[62753]: DEBUG oslo_vmware.api [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332228, 'name': RemoveSnapshot_Task, 'duration_secs': 1.280488} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 874.561888] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Deleted Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 874.561888] env[62753]: INFO nova.compute.manager [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Took 14.26 seconds to snapshot the instance on the hypervisor. [ 874.624177] env[62753]: DEBUG nova.scheduler.client.report [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 874.648928] env[62753]: DEBUG nova.network.neutron [-] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.753417] env[62753]: DEBUG nova.network.neutron [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Updating instance_info_cache with network_info: [{"id": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "address": "fa:16:3e:1d:84:1b", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8e502a-7a", "ovs_interfaceid": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 874.756722] env[62753]: DEBUG nova.virt.hardware [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.756921] env[62753]: DEBUG nova.virt.hardware [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.757096] env[62753]: DEBUG nova.virt.hardware [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.758027] env[62753]: DEBUG nova.virt.hardware [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.758027] env[62753]: DEBUG nova.virt.hardware [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.758027] env[62753]: DEBUG nova.virt.hardware [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.758027] env[62753]: DEBUG nova.virt.hardware [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.758228] env[62753]: DEBUG nova.virt.hardware [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.758333] env[62753]: DEBUG nova.virt.hardware [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.758509] env[62753]: DEBUG nova.virt.hardware [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.759118] env[62753]: DEBUG nova.virt.hardware [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.759705] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13bad09b-e479-4d2c-9a3f-0b3ed9f2bfe5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.767965] env[62753]: DEBUG nova.compute.manager [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 874.772579] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14dfa035-b24f-41b4-b312-e52c4e53749c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.776476] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 874.776689] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 874.776873] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Deleting the datastore file [datastore2] 046fc3dc-a187-4fda-bc66-345e1226f83d {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 874.777442] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c46686eb-e207-4908-827f-75bd314ad16e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.789753] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:98:33', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9d39252e-42ef-4252-98d3-62af5a0d109d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 874.797215] env[62753]: DEBUG oslo.service.loopingcall [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 874.800332] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 874.800948] env[62753]: DEBUG oslo_vmware.api [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 874.800948] env[62753]: value = "task-1332235" [ 874.800948] env[62753]: _type = "Task" [ 874.800948] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.804941] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b77dd2be-5962-490e-8d17-e3987a1e418e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.825239] env[62753]: DEBUG nova.virt.hardware [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 874.825570] env[62753]: DEBUG nova.virt.hardware [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 874.825824] env[62753]: DEBUG nova.virt.hardware [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 874.826134] env[62753]: DEBUG nova.virt.hardware [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 874.826419] env[62753]: DEBUG nova.virt.hardware [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 874.826699] env[62753]: DEBUG nova.virt.hardware [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 874.827226] env[62753]: DEBUG nova.virt.hardware [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 874.827695] env[62753]: DEBUG nova.virt.hardware [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 874.827695] env[62753]: DEBUG nova.virt.hardware [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 874.827853] env[62753]: DEBUG nova.virt.hardware [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 874.828141] env[62753]: DEBUG nova.virt.hardware [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 874.832523] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc2b1f2f-d886-4e54-b192-1e227a830115 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.840477] env[62753]: DEBUG oslo_vmware.api [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332235, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 874.843540] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 874.843540] env[62753]: value = "task-1332236" [ 874.843540] env[62753]: _type = "Task" [ 874.843540] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 874.844832] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-603eb3fc-fd39-47f8-ba30-f0932e1472e4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 874.855850] env[62753]: DEBUG nova.compute.manager [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 874.879198] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332236, 'name': CreateVM_Task} progress is 15%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.066086] env[62753]: DEBUG nova.compute.manager [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Instance disappeared during snapshot {{(pid=62753) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4494}} [ 875.080014] env[62753]: DEBUG nova.compute.manager [None req-a1015683-a8a4-4f0b-9cff-60a2b403ee66 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Image not found during clean up f136d03d-2911-4542-8963-164084c24bfe {{(pid=62753) _snapshot_instance /opt/stack/nova/nova/compute/manager.py:4500}} [ 875.131967] env[62753]: DEBUG oslo_concurrency.lockutils [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.390s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 875.131967] env[62753]: DEBUG nova.compute.manager [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 875.133594] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.200s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 875.139118] env[62753]: INFO nova.compute.claims [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 875.150866] env[62753]: INFO nova.compute.manager [-] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Took 1.02 seconds to deallocate network for instance. [ 875.263968] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Releasing lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.263968] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Updated the network info_cache for instance {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 875.332422] env[62753]: DEBUG oslo_vmware.api [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332235, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.171585} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.332721] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 875.332924] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 875.333134] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 875.333452] env[62753]: INFO nova.compute.manager [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Took 1.82 seconds to destroy the instance on the hypervisor. [ 875.333783] env[62753]: DEBUG oslo.service.loopingcall [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 875.334026] env[62753]: DEBUG nova.compute.manager [-] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 875.334120] env[62753]: DEBUG nova.network.neutron [-] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 875.340131] env[62753]: DEBUG nova.compute.manager [req-264f1fa0-9524-4916-ad56-182e027777f4 req-8ad9ed33-c3d7-4b45-b129-7c5af0ca2d55 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Received event network-changed-7fb046f4-e470-4f6d-8fa3-73f3b0157020 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.340870] env[62753]: DEBUG nova.compute.manager [req-264f1fa0-9524-4916-ad56-182e027777f4 req-8ad9ed33-c3d7-4b45-b129-7c5af0ca2d55 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Refreshing instance network info cache due to event network-changed-7fb046f4-e470-4f6d-8fa3-73f3b0157020. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 875.340870] env[62753]: DEBUG oslo_concurrency.lockutils [req-264f1fa0-9524-4916-ad56-182e027777f4 req-8ad9ed33-c3d7-4b45-b129-7c5af0ca2d55 service nova] Acquiring lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.340870] env[62753]: DEBUG oslo_concurrency.lockutils [req-264f1fa0-9524-4916-ad56-182e027777f4 req-8ad9ed33-c3d7-4b45-b129-7c5af0ca2d55 service nova] Acquired lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.340870] env[62753]: DEBUG nova.network.neutron [req-264f1fa0-9524-4916-ad56-182e027777f4 req-8ad9ed33-c3d7-4b45-b129-7c5af0ca2d55 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Refreshing network info cache for port 7fb046f4-e470-4f6d-8fa3-73f3b0157020 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 875.359287] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332236, 'name': CreateVM_Task, 'duration_secs': 0.491072} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.359548] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 875.360084] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.360316] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.360657] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 875.360814] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-00ec50db-9414-4b20-bcbd-c69ae04ec64e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.365287] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 875.365287] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522e1ea7-9015-4ca0-a01d-7b363739d6e9" [ 875.365287] env[62753]: _type = "Task" [ 875.365287] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.374996] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522e1ea7-9015-4ca0-a01d-7b363739d6e9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 875.394332] env[62753]: DEBUG oslo_concurrency.lockutils [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.614730] env[62753]: DEBUG nova.compute.manager [req-615cb8fd-0d24-459f-b526-85637f71210c req-263648fa-155f-44be-80ff-0f48636bf1ab service nova] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Received event network-vif-deleted-ef0eb4ad-a83b-435a-a813-4434a8616c14 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 875.615194] env[62753]: INFO nova.compute.manager [req-615cb8fd-0d24-459f-b526-85637f71210c req-263648fa-155f-44be-80ff-0f48636bf1ab service nova] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Neutron deleted interface ef0eb4ad-a83b-435a-a813-4434a8616c14; detaching it from the instance and deleting it from the info cache [ 875.615476] env[62753]: DEBUG nova.network.neutron [req-615cb8fd-0d24-459f-b526-85637f71210c req-263648fa-155f-44be-80ff-0f48636bf1ab service nova] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 875.647830] env[62753]: DEBUG nova.compute.utils [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 875.652915] env[62753]: DEBUG nova.compute.manager [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 875.653018] env[62753]: DEBUG nova.network.neutron [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 875.657658] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 875.707139] env[62753]: DEBUG nova.policy [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9f17c221bd924a648af7fb7ec495018b', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'a83c3627d2e040709b4d2d747758776b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 875.718731] env[62753]: DEBUG nova.network.neutron [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Successfully updated port: 2b4dd6f0-5b5b-47ee-b7cb-ecafe1190758 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 875.877702] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]522e1ea7-9015-4ca0-a01d-7b363739d6e9, 'name': SearchDatastore_Task, 'duration_secs': 0.009385} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 875.880316] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 875.880560] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 875.880834] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 875.881026] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 875.881220] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 875.881826] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-62056e38-c81c-4be0-998f-db5cb117cd7b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.892099] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 875.892410] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 875.893211] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b7055ba5-bb58-4ed9-bedc-ad60b922a635 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 875.898555] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 875.898555] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52de7af1-e969-8173-df25-6826c477ca1e" [ 875.898555] env[62753]: _type = "Task" [ 875.898555] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 875.908119] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52de7af1-e969-8173-df25-6826c477ca1e, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.035261] env[62753]: DEBUG nova.network.neutron [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Successfully created port: e44dcb1f-60ee-409a-805b-0c795852f967 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 876.078481] env[62753]: DEBUG nova.network.neutron [req-264f1fa0-9524-4916-ad56-182e027777f4 req-8ad9ed33-c3d7-4b45-b129-7c5af0ca2d55 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Updated VIF entry in instance network info cache for port 7fb046f4-e470-4f6d-8fa3-73f3b0157020. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 876.078918] env[62753]: DEBUG nova.network.neutron [req-264f1fa0-9524-4916-ad56-182e027777f4 req-8ad9ed33-c3d7-4b45-b129-7c5af0ca2d55 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Updating instance_info_cache with network_info: [{"id": "7fb046f4-e470-4f6d-8fa3-73f3b0157020", "address": "fa:16:3e:6f:26:2c", "network": {"id": "fc646f7f-70ba-4671-8476-3b0ece1e9041", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-853174775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d0eb0b3ba504f59a85cb021da2b47b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fb046f4-e4", "ovs_interfaceid": "7fb046f4-e470-4f6d-8fa3-73f3b0157020", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.091323] env[62753]: DEBUG nova.network.neutron [-] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 876.120361] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6eeeade1-25ea-4428-8e32-6082c2228370 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.131125] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6adfa1ab-d9a2-4f6a-b5ae-99dd170eff30 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.157666] env[62753]: DEBUG nova.compute.manager [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 876.158369] env[62753]: DEBUG nova.compute.manager [req-615cb8fd-0d24-459f-b526-85637f71210c req-263648fa-155f-44be-80ff-0f48636bf1ab service nova] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Detach interface failed, port_id=ef0eb4ad-a83b-435a-a813-4434a8616c14, reason: Instance 046fc3dc-a187-4fda-bc66-345e1226f83d could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 876.411291] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52de7af1-e969-8173-df25-6826c477ca1e, 'name': SearchDatastore_Task, 'duration_secs': 0.012148} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.412060] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-129a13b4-9136-4f8c-86d8-7780c5e42658 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.418671] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 876.418671] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529ccca7-ef85-20b0-bb1b-3abc2930e4ea" [ 876.418671] env[62753]: _type = "Task" [ 876.418671] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.426636] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529ccca7-ef85-20b0-bb1b-3abc2930e4ea, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 876.525567] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bce92a3-1c4f-4c1d-a01c-05011289af66 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.533198] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83daec5d-63df-478f-8eb0-65762f295e1a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.564731] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24dc3377-8379-4b77-824c-4b306d165f01 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.572378] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28457ea7-444e-4e18-9d2b-70c5417c2bf6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.586177] env[62753]: DEBUG oslo_concurrency.lockutils [req-264f1fa0-9524-4916-ad56-182e027777f4 req-8ad9ed33-c3d7-4b45-b129-7c5af0ca2d55 service nova] Releasing lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.586812] env[62753]: DEBUG nova.compute.provider_tree [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 876.594327] env[62753]: INFO nova.compute.manager [-] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Took 1.26 seconds to deallocate network for instance. [ 876.665848] env[62753]: INFO nova.virt.block_device [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Booting with volume 02056526-1ec1-42b1-96d8-e237f65687ce at /dev/sda [ 876.719439] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2a7f2c1e-1a9b-477c-ba71-0c862933f167 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.731963] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a5e7753-d557-44dc-b037-23aee81d5c0c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.765169] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-22fd8d4e-5163-4dfe-92a1-5156b2c5370c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.776393] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd0bde36-586c-41f9-88d0-78f894d4e67f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.811255] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-353bc812-ff6b-4357-a4e5-eeff6b49d0bb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.820980] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30555337-c9c9-4c5b-9f8b-302dc60cfcc5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.838773] env[62753]: DEBUG nova.virt.block_device [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Updating existing volume attachment record: 2af5fc9d-fed0-4fc7-8c57-e7a7d0048692 {{(pid=62753) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 876.931031] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529ccca7-ef85-20b0-bb1b-3abc2930e4ea, 'name': SearchDatastore_Task, 'duration_secs': 0.010598} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 876.931375] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 876.931640] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] a6ee010c-6305-4009-80e0-92a2c58bcd7b/a6ee010c-6305-4009-80e0-92a2c58bcd7b.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 876.932072] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a14c5da1-9dda-44f4-93ed-0b82003809c6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 876.939134] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 876.939134] env[62753]: value = "task-1332237" [ 876.939134] env[62753]: _type = "Task" [ 876.939134] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 876.949751] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332237, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.090286] env[62753]: DEBUG nova.scheduler.client.report [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 877.101828] env[62753]: DEBUG oslo_concurrency.lockutils [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.370788] env[62753]: DEBUG nova.compute.manager [req-efcfe4a7-412e-4fc3-8752-05020a247b25 req-e848df44-3c27-46da-ad1f-acb67aa1276c service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Received event network-vif-plugged-2b4dd6f0-5b5b-47ee-b7cb-ecafe1190758 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.371116] env[62753]: DEBUG oslo_concurrency.lockutils [req-efcfe4a7-412e-4fc3-8752-05020a247b25 req-e848df44-3c27-46da-ad1f-acb67aa1276c service nova] Acquiring lock "49311ef5-733a-4146-87ac-876a121a8d4d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.371239] env[62753]: DEBUG oslo_concurrency.lockutils [req-efcfe4a7-412e-4fc3-8752-05020a247b25 req-e848df44-3c27-46da-ad1f-acb67aa1276c service nova] Lock "49311ef5-733a-4146-87ac-876a121a8d4d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.371415] env[62753]: DEBUG oslo_concurrency.lockutils [req-efcfe4a7-412e-4fc3-8752-05020a247b25 req-e848df44-3c27-46da-ad1f-acb67aa1276c service nova] Lock "49311ef5-733a-4146-87ac-876a121a8d4d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.371606] env[62753]: DEBUG nova.compute.manager [req-efcfe4a7-412e-4fc3-8752-05020a247b25 req-e848df44-3c27-46da-ad1f-acb67aa1276c service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] No waiting events found dispatching network-vif-plugged-2b4dd6f0-5b5b-47ee-b7cb-ecafe1190758 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 877.371752] env[62753]: WARNING nova.compute.manager [req-efcfe4a7-412e-4fc3-8752-05020a247b25 req-e848df44-3c27-46da-ad1f-acb67aa1276c service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Received unexpected event network-vif-plugged-2b4dd6f0-5b5b-47ee-b7cb-ecafe1190758 for instance with vm_state building and task_state spawning. [ 877.372103] env[62753]: DEBUG nova.compute.manager [req-efcfe4a7-412e-4fc3-8752-05020a247b25 req-e848df44-3c27-46da-ad1f-acb67aa1276c service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Received event network-changed-2b4dd6f0-5b5b-47ee-b7cb-ecafe1190758 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.372277] env[62753]: DEBUG nova.compute.manager [req-efcfe4a7-412e-4fc3-8752-05020a247b25 req-e848df44-3c27-46da-ad1f-acb67aa1276c service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Refreshing instance network info cache due to event network-changed-2b4dd6f0-5b5b-47ee-b7cb-ecafe1190758. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 877.372472] env[62753]: DEBUG oslo_concurrency.lockutils [req-efcfe4a7-412e-4fc3-8752-05020a247b25 req-e848df44-3c27-46da-ad1f-acb67aa1276c service nova] Acquiring lock "refresh_cache-49311ef5-733a-4146-87ac-876a121a8d4d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 877.372614] env[62753]: DEBUG oslo_concurrency.lockutils [req-efcfe4a7-412e-4fc3-8752-05020a247b25 req-e848df44-3c27-46da-ad1f-acb67aa1276c service nova] Acquired lock "refresh_cache-49311ef5-733a-4146-87ac-876a121a8d4d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 877.372812] env[62753]: DEBUG nova.network.neutron [req-efcfe4a7-412e-4fc3-8752-05020a247b25 req-e848df44-3c27-46da-ad1f-acb67aa1276c service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Refreshing network info cache for port 2b4dd6f0-5b5b-47ee-b7cb-ecafe1190758 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 877.453819] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332237, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 877.596167] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.462s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.600022] env[62753]: DEBUG nova.compute.manager [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 877.601511] env[62753]: DEBUG oslo_concurrency.lockutils [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.823s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.604272] env[62753]: INFO nova.compute.claims [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 877.804115] env[62753]: DEBUG nova.network.neutron [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Successfully updated port: 51f3d79a-315b-418b-ac51-a3b5f5c328f3 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 877.930108] env[62753]: DEBUG nova.compute.manager [req-275ca18b-6862-4126-bbd6-30c991ec2191 req-e30267a5-5b3d-408f-803a-8cd1ad7a18cf service nova] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Received event network-vif-plugged-e44dcb1f-60ee-409a-805b-0c795852f967 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 877.930515] env[62753]: DEBUG oslo_concurrency.lockutils [req-275ca18b-6862-4126-bbd6-30c991ec2191 req-e30267a5-5b3d-408f-803a-8cd1ad7a18cf service nova] Acquiring lock "c5dae9a3-052b-4f4c-ac79-84aeaded457a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 877.934244] env[62753]: DEBUG oslo_concurrency.lockutils [req-275ca18b-6862-4126-bbd6-30c991ec2191 req-e30267a5-5b3d-408f-803a-8cd1ad7a18cf service nova] Lock "c5dae9a3-052b-4f4c-ac79-84aeaded457a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 877.934244] env[62753]: DEBUG oslo_concurrency.lockutils [req-275ca18b-6862-4126-bbd6-30c991ec2191 req-e30267a5-5b3d-408f-803a-8cd1ad7a18cf service nova] Lock "c5dae9a3-052b-4f4c-ac79-84aeaded457a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 877.934244] env[62753]: DEBUG nova.compute.manager [req-275ca18b-6862-4126-bbd6-30c991ec2191 req-e30267a5-5b3d-408f-803a-8cd1ad7a18cf service nova] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] No waiting events found dispatching network-vif-plugged-e44dcb1f-60ee-409a-805b-0c795852f967 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 877.934244] env[62753]: WARNING nova.compute.manager [req-275ca18b-6862-4126-bbd6-30c991ec2191 req-e30267a5-5b3d-408f-803a-8cd1ad7a18cf service nova] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Received unexpected event network-vif-plugged-e44dcb1f-60ee-409a-805b-0c795852f967 for instance with vm_state building and task_state block_device_mapping. [ 877.941186] env[62753]: DEBUG nova.network.neutron [req-efcfe4a7-412e-4fc3-8752-05020a247b25 req-e848df44-3c27-46da-ad1f-acb67aa1276c service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 877.951899] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332237, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.718237} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 877.952699] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] a6ee010c-6305-4009-80e0-92a2c58bcd7b/a6ee010c-6305-4009-80e0-92a2c58bcd7b.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 877.953031] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 877.953271] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d1349567-fe57-4ccb-92c3-fb4b98eb49df {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 877.961500] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 877.961500] env[62753]: value = "task-1332238" [ 877.961500] env[62753]: _type = "Task" [ 877.961500] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 877.978038] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332238, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.043383] env[62753]: DEBUG nova.network.neutron [req-efcfe4a7-412e-4fc3-8752-05020a247b25 req-e848df44-3c27-46da-ad1f-acb67aa1276c service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 878.108407] env[62753]: DEBUG nova.compute.utils [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 878.111795] env[62753]: DEBUG nova.compute.manager [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 878.111795] env[62753]: DEBUG nova.network.neutron [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 878.162467] env[62753]: DEBUG nova.policy [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6ec6b26d1d5444e93c410ee4e2ea806', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '371a5e8b16be4cffbdeb2139757684fc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 878.356129] env[62753]: DEBUG nova.network.neutron [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Successfully updated port: e44dcb1f-60ee-409a-805b-0c795852f967 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 878.392056] env[62753]: DEBUG nova.compute.manager [req-b1548ca7-341f-4f81-a759-ff1fa094774d req-b721d481-055e-4fcd-b3cb-629c517dcf33 service nova] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Received event network-changed-e44dcb1f-60ee-409a-805b-0c795852f967 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 878.392266] env[62753]: DEBUG nova.compute.manager [req-b1548ca7-341f-4f81-a759-ff1fa094774d req-b721d481-055e-4fcd-b3cb-629c517dcf33 service nova] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Refreshing instance network info cache due to event network-changed-e44dcb1f-60ee-409a-805b-0c795852f967. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 878.392593] env[62753]: DEBUG oslo_concurrency.lockutils [req-b1548ca7-341f-4f81-a759-ff1fa094774d req-b721d481-055e-4fcd-b3cb-629c517dcf33 service nova] Acquiring lock "refresh_cache-c5dae9a3-052b-4f4c-ac79-84aeaded457a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.392813] env[62753]: DEBUG oslo_concurrency.lockutils [req-b1548ca7-341f-4f81-a759-ff1fa094774d req-b721d481-055e-4fcd-b3cb-629c517dcf33 service nova] Acquired lock "refresh_cache-c5dae9a3-052b-4f4c-ac79-84aeaded457a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 878.392990] env[62753]: DEBUG nova.network.neutron [req-b1548ca7-341f-4f81-a759-ff1fa094774d req-b721d481-055e-4fcd-b3cb-629c517dcf33 service nova] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Refreshing network info cache for port e44dcb1f-60ee-409a-805b-0c795852f967 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 878.473526] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332238, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075124} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 878.473809] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 878.474583] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8946a3d4-0230-46ec-b05c-5248eb16afd2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.496468] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Reconfiguring VM instance instance-0000003d to attach disk [datastore1] a6ee010c-6305-4009-80e0-92a2c58bcd7b/a6ee010c-6305-4009-80e0-92a2c58bcd7b.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 878.496831] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-63de8eae-e8e7-4a90-b088-6951e0f037a5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.521528] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 878.521528] env[62753]: value = "task-1332239" [ 878.521528] env[62753]: _type = "Task" [ 878.521528] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 878.532464] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332239, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 878.546527] env[62753]: DEBUG oslo_concurrency.lockutils [req-efcfe4a7-412e-4fc3-8752-05020a247b25 req-e848df44-3c27-46da-ad1f-acb67aa1276c service nova] Releasing lock "refresh_cache-49311ef5-733a-4146-87ac-876a121a8d4d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 878.548384] env[62753]: DEBUG nova.network.neutron [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Successfully created port: 9c52b1e4-82d2-4449-91f3-a624596bdc79 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 878.612356] env[62753]: DEBUG nova.compute.manager [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 878.858818] env[62753]: DEBUG oslo_concurrency.lockutils [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Acquiring lock "refresh_cache-c5dae9a3-052b-4f4c-ac79-84aeaded457a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 878.893250] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ad605c-bb17-454a-ae5f-38d83d05ec23 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.902269] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d63cfb-1fc5-471f-9d55-7afec51c32fd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.934425] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e9aa9d9-1bac-46ef-81b3-65ce305f8f9e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.937641] env[62753]: DEBUG nova.network.neutron [req-b1548ca7-341f-4f81-a759-ff1fa094774d req-b721d481-055e-4fcd-b3cb-629c517dcf33 service nova] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 878.944039] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95054321-d522-4cbf-b690-afb2e2be02db {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.956985] env[62753]: DEBUG nova.compute.provider_tree [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 878.969641] env[62753]: DEBUG nova.compute.manager [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 878.970155] env[62753]: DEBUG nova.virt.hardware [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 878.970372] env[62753]: DEBUG nova.virt.hardware [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 878.970531] env[62753]: DEBUG nova.virt.hardware [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 878.970712] env[62753]: DEBUG nova.virt.hardware [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 878.970861] env[62753]: DEBUG nova.virt.hardware [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 878.971110] env[62753]: DEBUG nova.virt.hardware [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 878.971228] env[62753]: DEBUG nova.virt.hardware [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 878.971386] env[62753]: DEBUG nova.virt.hardware [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 878.971555] env[62753]: DEBUG nova.virt.hardware [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 878.971719] env[62753]: DEBUG nova.virt.hardware [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 878.972013] env[62753]: DEBUG nova.virt.hardware [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 878.973009] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-926d4867-f618-4690-ad4b-87292291e127 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 878.980277] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d99f4ed4-a3da-4b30-811b-9b0baef0dcb5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.033367] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.069197] env[62753]: DEBUG nova.network.neutron [req-b1548ca7-341f-4f81-a759-ff1fa094774d req-b721d481-055e-4fcd-b3cb-629c517dcf33 service nova] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 879.396690] env[62753]: DEBUG nova.compute.manager [req-6c885cb0-57d2-4b96-b0b4-d256a25a3663 req-8b7a4bfc-f8bd-4bd3-a18d-1898ccc1592c service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Received event network-vif-plugged-51f3d79a-315b-418b-ac51-a3b5f5c328f3 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.396920] env[62753]: DEBUG oslo_concurrency.lockutils [req-6c885cb0-57d2-4b96-b0b4-d256a25a3663 req-8b7a4bfc-f8bd-4bd3-a18d-1898ccc1592c service nova] Acquiring lock "49311ef5-733a-4146-87ac-876a121a8d4d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 879.397159] env[62753]: DEBUG oslo_concurrency.lockutils [req-6c885cb0-57d2-4b96-b0b4-d256a25a3663 req-8b7a4bfc-f8bd-4bd3-a18d-1898ccc1592c service nova] Lock "49311ef5-733a-4146-87ac-876a121a8d4d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.397340] env[62753]: DEBUG oslo_concurrency.lockutils [req-6c885cb0-57d2-4b96-b0b4-d256a25a3663 req-8b7a4bfc-f8bd-4bd3-a18d-1898ccc1592c service nova] Lock "49311ef5-733a-4146-87ac-876a121a8d4d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.397513] env[62753]: DEBUG nova.compute.manager [req-6c885cb0-57d2-4b96-b0b4-d256a25a3663 req-8b7a4bfc-f8bd-4bd3-a18d-1898ccc1592c service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] No waiting events found dispatching network-vif-plugged-51f3d79a-315b-418b-ac51-a3b5f5c328f3 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 879.397683] env[62753]: WARNING nova.compute.manager [req-6c885cb0-57d2-4b96-b0b4-d256a25a3663 req-8b7a4bfc-f8bd-4bd3-a18d-1898ccc1592c service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Received unexpected event network-vif-plugged-51f3d79a-315b-418b-ac51-a3b5f5c328f3 for instance with vm_state building and task_state spawning. [ 879.397849] env[62753]: DEBUG nova.compute.manager [req-6c885cb0-57d2-4b96-b0b4-d256a25a3663 req-8b7a4bfc-f8bd-4bd3-a18d-1898ccc1592c service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Received event network-changed-51f3d79a-315b-418b-ac51-a3b5f5c328f3 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 879.398017] env[62753]: DEBUG nova.compute.manager [req-6c885cb0-57d2-4b96-b0b4-d256a25a3663 req-8b7a4bfc-f8bd-4bd3-a18d-1898ccc1592c service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Refreshing instance network info cache due to event network-changed-51f3d79a-315b-418b-ac51-a3b5f5c328f3. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 879.398215] env[62753]: DEBUG oslo_concurrency.lockutils [req-6c885cb0-57d2-4b96-b0b4-d256a25a3663 req-8b7a4bfc-f8bd-4bd3-a18d-1898ccc1592c service nova] Acquiring lock "refresh_cache-49311ef5-733a-4146-87ac-876a121a8d4d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 879.398358] env[62753]: DEBUG oslo_concurrency.lockutils [req-6c885cb0-57d2-4b96-b0b4-d256a25a3663 req-8b7a4bfc-f8bd-4bd3-a18d-1898ccc1592c service nova] Acquired lock "refresh_cache-49311ef5-733a-4146-87ac-876a121a8d4d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.398561] env[62753]: DEBUG nova.network.neutron [req-6c885cb0-57d2-4b96-b0b4-d256a25a3663 req-8b7a4bfc-f8bd-4bd3-a18d-1898ccc1592c service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Refreshing network info cache for port 51f3d79a-315b-418b-ac51-a3b5f5c328f3 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 879.460026] env[62753]: DEBUG nova.scheduler.client.report [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 879.534163] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332239, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 879.571342] env[62753]: DEBUG oslo_concurrency.lockutils [req-b1548ca7-341f-4f81-a759-ff1fa094774d req-b721d481-055e-4fcd-b3cb-629c517dcf33 service nova] Releasing lock "refresh_cache-c5dae9a3-052b-4f4c-ac79-84aeaded457a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 879.571713] env[62753]: DEBUG oslo_concurrency.lockutils [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Acquired lock "refresh_cache-c5dae9a3-052b-4f4c-ac79-84aeaded457a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 879.571923] env[62753]: DEBUG nova.network.neutron [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 879.624280] env[62753]: DEBUG nova.compute.manager [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 879.651399] env[62753]: DEBUG nova.virt.hardware [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 879.651660] env[62753]: DEBUG nova.virt.hardware [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 879.651849] env[62753]: DEBUG nova.virt.hardware [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 879.652036] env[62753]: DEBUG nova.virt.hardware [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 879.652190] env[62753]: DEBUG nova.virt.hardware [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 879.652339] env[62753]: DEBUG nova.virt.hardware [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 879.652553] env[62753]: DEBUG nova.virt.hardware [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 879.652773] env[62753]: DEBUG nova.virt.hardware [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 879.652930] env[62753]: DEBUG nova.virt.hardware [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 879.653116] env[62753]: DEBUG nova.virt.hardware [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 879.653299] env[62753]: DEBUG nova.virt.hardware [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 879.654171] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-023f72a1-513f-4629-a97d-058d239e9a33 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.661774] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb03f2b0-85e9-4c1a-a332-d2db32997a6a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 879.954536] env[62753]: DEBUG nova.network.neutron [req-6c885cb0-57d2-4b96-b0b4-d256a25a3663 req-8b7a4bfc-f8bd-4bd3-a18d-1898ccc1592c service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 879.964930] env[62753]: DEBUG oslo_concurrency.lockutils [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.363s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 879.965679] env[62753]: DEBUG nova.compute.manager [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 879.968561] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 23.477s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 879.972225] env[62753]: INFO nova.compute.claims [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 880.040831] env[62753]: DEBUG nova.network.neutron [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Successfully updated port: f07bc1fd-c64f-4525-b198-0d7cc81d6927 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 880.045578] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332239, 'name': ReconfigVM_Task, 'duration_secs': 1.421558} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.045578] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Reconfigured VM instance instance-0000003d to attach disk [datastore1] a6ee010c-6305-4009-80e0-92a2c58bcd7b/a6ee010c-6305-4009-80e0-92a2c58bcd7b.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 880.045578] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-18dcecad-05b0-42bc-aff3-0d3423328a8a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.051690] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 880.051690] env[62753]: value = "task-1332240" [ 880.051690] env[62753]: _type = "Task" [ 880.051690] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.062851] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332240, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.075683] env[62753]: DEBUG nova.network.neutron [req-6c885cb0-57d2-4b96-b0b4-d256a25a3663 req-8b7a4bfc-f8bd-4bd3-a18d-1898ccc1592c service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.127478] env[62753]: DEBUG nova.network.neutron [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 880.212588] env[62753]: DEBUG nova.network.neutron [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Successfully updated port: 9c52b1e4-82d2-4449-91f3-a624596bdc79 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 880.311503] env[62753]: DEBUG nova.network.neutron [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Updating instance_info_cache with network_info: [{"id": "e44dcb1f-60ee-409a-805b-0c795852f967", "address": "fa:16:3e:0d:48:c3", "network": {"id": "c07a7726-4e7d-4f78-bdc8-ca9039c7997f", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-687128365-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a83c3627d2e040709b4d2d747758776b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8bb798a9-4c7f-4361-9436-a5690283861a", "external-id": "nsx-vlan-transportzone-547", "segmentation_id": 547, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape44dcb1f-60", "ovs_interfaceid": "e44dcb1f-60ee-409a-805b-0c795852f967", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 880.476610] env[62753]: DEBUG nova.compute.utils [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 880.479975] env[62753]: DEBUG nova.compute.manager [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Not allocating networking since 'none' was specified. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 880.545227] env[62753]: DEBUG oslo_concurrency.lockutils [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquiring lock "refresh_cache-49311ef5-733a-4146-87ac-876a121a8d4d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.564171] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332240, 'name': Rename_Task, 'duration_secs': 0.210093} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 880.564500] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 880.564778] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9dffe358-54d0-4c5f-a74b-58bf4f6fd8af {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.571607] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 880.571607] env[62753]: value = "task-1332241" [ 880.571607] env[62753]: _type = "Task" [ 880.571607] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.580523] env[62753]: DEBUG oslo_concurrency.lockutils [req-6c885cb0-57d2-4b96-b0b4-d256a25a3663 req-8b7a4bfc-f8bd-4bd3-a18d-1898ccc1592c service nova] Releasing lock "refresh_cache-49311ef5-733a-4146-87ac-876a121a8d4d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.580890] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332241, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.581112] env[62753]: DEBUG oslo_concurrency.lockutils [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquired lock "refresh_cache-49311ef5-733a-4146-87ac-876a121a8d4d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.581269] env[62753]: DEBUG nova.network.neutron [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 880.715972] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 880.716144] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquired lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 880.716303] env[62753]: DEBUG nova.network.neutron [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 880.814254] env[62753]: DEBUG oslo_concurrency.lockutils [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Releasing lock "refresh_cache-c5dae9a3-052b-4f4c-ac79-84aeaded457a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 880.814585] env[62753]: DEBUG nova.compute.manager [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Instance network_info: |[{"id": "e44dcb1f-60ee-409a-805b-0c795852f967", "address": "fa:16:3e:0d:48:c3", "network": {"id": "c07a7726-4e7d-4f78-bdc8-ca9039c7997f", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-687128365-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a83c3627d2e040709b4d2d747758776b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8bb798a9-4c7f-4361-9436-a5690283861a", "external-id": "nsx-vlan-transportzone-547", "segmentation_id": 547, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape44dcb1f-60", "ovs_interfaceid": "e44dcb1f-60ee-409a-805b-0c795852f967", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 880.815110] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0d:48:c3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8bb798a9-4c7f-4361-9436-a5690283861a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'e44dcb1f-60ee-409a-805b-0c795852f967', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 880.822454] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Creating folder: Project (a83c3627d2e040709b4d2d747758776b). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 880.822753] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f2be94bc-bb88-4eb2-899e-a6e24c9fc72d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.836587] env[62753]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 880.836750] env[62753]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62753) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 880.837374] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Folder already exists: Project (a83c3627d2e040709b4d2d747758776b). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 880.837588] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Creating folder: Instances. Parent ref: group-v284555. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 880.837829] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-070169b5-08a9-48bb-be52-bf7a7058c15d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.847599] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Created folder: Instances in parent group-v284555. [ 880.847845] env[62753]: DEBUG oslo.service.loopingcall [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 880.848040] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 880.848248] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e0dc1b2a-a95a-45f7-8794-c3930ebb59f3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 880.866961] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 880.866961] env[62753]: value = "task-1332244" [ 880.866961] env[62753]: _type = "Task" [ 880.866961] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 880.874268] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332244, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 880.980975] env[62753]: DEBUG nova.compute.manager [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 881.081201] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332241, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.143869] env[62753]: DEBUG nova.network.neutron [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 881.259207] env[62753]: DEBUG nova.network.neutron [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 881.262424] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50662be0-21ff-4a20-86ed-ea702646170a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.270554] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9873ccac-60da-42e5-9cc8-7ec505236710 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.302741] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70d9703d-638d-4998-bb21-47e7cd10c4e8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.310255] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21009178-9c91-4ca3-93eb-3cc1e58ae523 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.325242] env[62753]: DEBUG nova.compute.provider_tree [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 881.377996] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332244, 'name': CreateVM_Task, 'duration_secs': 0.35409} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.378193] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 881.378867] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'guest_format': None, 'delete_on_termination': True, 'attachment_id': '2af5fc9d-fed0-4fc7-8c57-e7a7d0048692', 'device_type': None, 'mount_device': '/dev/sda', 'disk_bus': None, 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284559', 'volume_id': '02056526-1ec1-42b1-96d8-e237f65687ce', 'name': 'volume-02056526-1ec1-42b1-96d8-e237f65687ce', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c5dae9a3-052b-4f4c-ac79-84aeaded457a', 'attached_at': '', 'detached_at': '', 'volume_id': '02056526-1ec1-42b1-96d8-e237f65687ce', 'serial': '02056526-1ec1-42b1-96d8-e237f65687ce'}, 'volume_type': None}], 'swap': None} {{(pid=62753) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 881.379106] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Root volume attach. Driver type: vmdk {{(pid=62753) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 881.380131] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-329d404f-d368-467a-81f3-47affdca6dc2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.388615] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe29ea51-b1e9-4e28-9033-2d9e51c50920 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.394319] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-642505b3-e168-4120-90b6-d9c54761e954 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.403591] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-dcde4e5b-d037-44b5-9309-68318585e6c7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.410508] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Waiting for the task: (returnval){ [ 881.410508] env[62753]: value = "task-1332245" [ 881.410508] env[62753]: _type = "Task" [ 881.410508] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 881.417853] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332245, 'name': RelocateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.427895] env[62753]: DEBUG nova.compute.manager [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Received event network-vif-plugged-f07bc1fd-c64f-4525-b198-0d7cc81d6927 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.428161] env[62753]: DEBUG oslo_concurrency.lockutils [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] Acquiring lock "49311ef5-733a-4146-87ac-876a121a8d4d-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 881.428655] env[62753]: DEBUG oslo_concurrency.lockutils [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] Lock "49311ef5-733a-4146-87ac-876a121a8d4d-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 881.428655] env[62753]: DEBUG oslo_concurrency.lockutils [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] Lock "49311ef5-733a-4146-87ac-876a121a8d4d-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 881.428793] env[62753]: DEBUG nova.compute.manager [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] No waiting events found dispatching network-vif-plugged-f07bc1fd-c64f-4525-b198-0d7cc81d6927 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 881.429078] env[62753]: WARNING nova.compute.manager [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Received unexpected event network-vif-plugged-f07bc1fd-c64f-4525-b198-0d7cc81d6927 for instance with vm_state building and task_state spawning. [ 881.429255] env[62753]: DEBUG nova.compute.manager [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Received event network-changed-f07bc1fd-c64f-4525-b198-0d7cc81d6927 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 881.429514] env[62753]: DEBUG nova.compute.manager [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Refreshing instance network info cache due to event network-changed-f07bc1fd-c64f-4525-b198-0d7cc81d6927. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 881.429725] env[62753]: DEBUG oslo_concurrency.lockutils [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] Acquiring lock "refresh_cache-49311ef5-733a-4146-87ac-876a121a8d4d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 881.435020] env[62753]: DEBUG nova.network.neutron [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Updating instance_info_cache with network_info: [{"id": "9c52b1e4-82d2-4449-91f3-a624596bdc79", "address": "fa:16:3e:8e:1b:de", "network": {"id": "12446a5e-7329-4f8f-aa28-a8b9d78e7573", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1766107072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "371a5e8b16be4cffbdeb2139757684fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7c7f16b-a2e3-4d1f-9b7e-a44a9ebb589c", "external-id": "nsx-vlan-transportzone-772", "segmentation_id": 772, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c52b1e4-82", "ovs_interfaceid": "9c52b1e4-82d2-4449-91f3-a624596bdc79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.585179] env[62753]: DEBUG oslo_vmware.api [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332241, 'name': PowerOnVM_Task, 'duration_secs': 0.631772} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 881.585179] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 881.585460] env[62753]: DEBUG nova.compute.manager [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 881.586219] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bccd814-dc27-40ed-922f-e68f915d95b2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.655787] env[62753]: DEBUG nova.network.neutron [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Updating instance_info_cache with network_info: [{"id": "2b4dd6f0-5b5b-47ee-b7cb-ecafe1190758", "address": "fa:16:3e:8f:a0:ec", "network": {"id": "ebb4962b-34ac-499f-afea-8131f81571e1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-220454372", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.149", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b01e80a17e843a98ee662157230ae17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b4dd6f0-5b", "ovs_interfaceid": "2b4dd6f0-5b5b-47ee-b7cb-ecafe1190758", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "51f3d79a-315b-418b-ac51-a3b5f5c328f3", "address": "fa:16:3e:2c:06:61", "network": {"id": "3e5a3d6d-6c74-4d41-91b4-136e08b33e52", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1495343464", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4b01e80a17e843a98ee662157230ae17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d2e4070-a78e-4d08-a104-b6312ab65577", "external-id": "nsx-vlan-transportzone-292", "segmentation_id": 292, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51f3d79a-31", "ovs_interfaceid": "51f3d79a-315b-418b-ac51-a3b5f5c328f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f07bc1fd-c64f-4525-b198-0d7cc81d6927", "address": "fa:16:3e:13:eb:0e", "network": {"id": "ebb4962b-34ac-499f-afea-8131f81571e1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-220454372", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.207", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b01e80a17e843a98ee662157230ae17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf07bc1fd-c6", "ovs_interfaceid": "f07bc1fd-c64f-4525-b198-0d7cc81d6927", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 881.830345] env[62753]: DEBUG nova.scheduler.client.report [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 881.923144] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332245, 'name': RelocateVM_Task} progress is 43%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 881.937526] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Releasing lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 881.937682] env[62753]: DEBUG nova.compute.manager [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Instance network_info: |[{"id": "9c52b1e4-82d2-4449-91f3-a624596bdc79", "address": "fa:16:3e:8e:1b:de", "network": {"id": "12446a5e-7329-4f8f-aa28-a8b9d78e7573", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1766107072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "371a5e8b16be4cffbdeb2139757684fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7c7f16b-a2e3-4d1f-9b7e-a44a9ebb589c", "external-id": "nsx-vlan-transportzone-772", "segmentation_id": 772, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c52b1e4-82", "ovs_interfaceid": "9c52b1e4-82d2-4449-91f3-a624596bdc79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 881.938106] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:1b:de', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c7c7f16b-a2e3-4d1f-9b7e-a44a9ebb589c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9c52b1e4-82d2-4449-91f3-a624596bdc79', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 881.945862] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Creating folder: Project (371a5e8b16be4cffbdeb2139757684fc). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 881.946568] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0fb35f3f-ef5e-4aab-9330-c7bdfa6d55f4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.958145] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Created folder: Project (371a5e8b16be4cffbdeb2139757684fc) in parent group-v284541. [ 881.958410] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Creating folder: Instances. Parent ref: group-v284585. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 881.958685] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-f263899e-1331-401c-adc1-e4f55938be12 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.970238] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Created folder: Instances in parent group-v284585. [ 881.970508] env[62753]: DEBUG oslo.service.loopingcall [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 881.970799] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d20e083-2959-453a-8875-47955bc02613] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 881.971056] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6e38016b-199f-4c8a-ac1d-89369dbeb80c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 881.992325] env[62753]: DEBUG nova.compute.manager [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 881.995900] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 881.995900] env[62753]: value = "task-1332248" [ 881.995900] env[62753]: _type = "Task" [ 881.995900] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.007777] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332248, 'name': CreateVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.020977] env[62753]: DEBUG nova.virt.hardware [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 882.020977] env[62753]: DEBUG nova.virt.hardware [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 882.020977] env[62753]: DEBUG nova.virt.hardware [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 882.021312] env[62753]: DEBUG nova.virt.hardware [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 882.021312] env[62753]: DEBUG nova.virt.hardware [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 882.021451] env[62753]: DEBUG nova.virt.hardware [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 882.021667] env[62753]: DEBUG nova.virt.hardware [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 882.021832] env[62753]: DEBUG nova.virt.hardware [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 882.022018] env[62753]: DEBUG nova.virt.hardware [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 882.022188] env[62753]: DEBUG nova.virt.hardware [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 882.022367] env[62753]: DEBUG nova.virt.hardware [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 882.023288] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a7825ff-4699-45ee-8521-d1fbfe06be69 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.033156] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f969c383-16e1-4f59-b59e-56ffc0327191 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.054299] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Instance VIF info [] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 882.061388] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Creating folder: Project (57176cb881784cbba124e2dc3e47ad6e). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 882.061820] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d1f1c871-02fd-4654-9409-3663c7c1f7a3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.073571] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Created folder: Project (57176cb881784cbba124e2dc3e47ad6e) in parent group-v284541. [ 882.073801] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Creating folder: Instances. Parent ref: group-v284588. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 882.074078] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a81d0928-5496-42e5-af02-c4658bba8325 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.084822] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Created folder: Instances in parent group-v284588. [ 882.085140] env[62753]: DEBUG oslo.service.loopingcall [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.085369] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 882.085589] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-08a38a03-5db4-4006-abac-8f6238461e9a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.104486] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 882.108119] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 882.108119] env[62753]: value = "task-1332251" [ 882.108119] env[62753]: _type = "Task" [ 882.108119] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.118800] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332251, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.158803] env[62753]: DEBUG oslo_concurrency.lockutils [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Releasing lock "refresh_cache-49311ef5-733a-4146-87ac-876a121a8d4d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 882.159385] env[62753]: DEBUG nova.compute.manager [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Instance network_info: |[{"id": "2b4dd6f0-5b5b-47ee-b7cb-ecafe1190758", "address": "fa:16:3e:8f:a0:ec", "network": {"id": "ebb4962b-34ac-499f-afea-8131f81571e1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-220454372", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.149", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b01e80a17e843a98ee662157230ae17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b4dd6f0-5b", "ovs_interfaceid": "2b4dd6f0-5b5b-47ee-b7cb-ecafe1190758", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "51f3d79a-315b-418b-ac51-a3b5f5c328f3", "address": "fa:16:3e:2c:06:61", "network": {"id": "3e5a3d6d-6c74-4d41-91b4-136e08b33e52", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1495343464", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4b01e80a17e843a98ee662157230ae17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d2e4070-a78e-4d08-a104-b6312ab65577", "external-id": "nsx-vlan-transportzone-292", "segmentation_id": 292, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51f3d79a-31", "ovs_interfaceid": "51f3d79a-315b-418b-ac51-a3b5f5c328f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f07bc1fd-c64f-4525-b198-0d7cc81d6927", "address": "fa:16:3e:13:eb:0e", "network": {"id": "ebb4962b-34ac-499f-afea-8131f81571e1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-220454372", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.207", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b01e80a17e843a98ee662157230ae17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf07bc1fd-c6", "ovs_interfaceid": "f07bc1fd-c64f-4525-b198-0d7cc81d6927", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 882.159747] env[62753]: DEBUG oslo_concurrency.lockutils [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] Acquired lock "refresh_cache-49311ef5-733a-4146-87ac-876a121a8d4d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.159970] env[62753]: DEBUG nova.network.neutron [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Refreshing network info cache for port f07bc1fd-c64f-4525-b198-0d7cc81d6927 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 882.163035] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8f:a0:ec', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c24464bb-bb6b-43a2-bdcd-8086ad1a307f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2b4dd6f0-5b5b-47ee-b7cb-ecafe1190758', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:2c:06:61', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '3d2e4070-a78e-4d08-a104-b6312ab65577', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '51f3d79a-315b-418b-ac51-a3b5f5c328f3', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:13:eb:0e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c24464bb-bb6b-43a2-bdcd-8086ad1a307f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f07bc1fd-c64f-4525-b198-0d7cc81d6927', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 882.174170] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Creating folder: Project (4b01e80a17e843a98ee662157230ae17). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 882.178213] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-952147e2-73fc-4493-9c28-666a3037fe05 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.191743] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Created folder: Project (4b01e80a17e843a98ee662157230ae17) in parent group-v284541. [ 882.192055] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Creating folder: Instances. Parent ref: group-v284591. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 882.192284] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-a58ef423-f142-4b26-9b7f-254e637ed19a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.208328] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Created folder: Instances in parent group-v284591. [ 882.208328] env[62753]: DEBUG oslo.service.loopingcall [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 882.208328] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 882.208328] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-45e7c1e8-c919-4e46-9267-6e62dfaf58dd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.235079] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 882.235079] env[62753]: value = "task-1332254" [ 882.235079] env[62753]: _type = "Task" [ 882.235079] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.248473] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332254, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.335746] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.367s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 882.336335] env[62753]: DEBUG nova.compute.manager [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 882.339215] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 22.297s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 882.342706] env[62753]: INFO nova.compute.claims [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 882.423076] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332245, 'name': RelocateVM_Task} progress is 56%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.507929] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332248, 'name': CreateVM_Task, 'duration_secs': 0.458518} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.510772] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d20e083-2959-453a-8875-47955bc02613] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 882.511665] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.511947] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 882.512311] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 882.512928] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c5062eed-27b6-4a18-8904-3b9c319ed5b0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 882.523768] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 882.523768] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a16416-1478-1ada-7bde-928fd7cecc77" [ 882.523768] env[62753]: _type = "Task" [ 882.523768] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 882.531953] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a16416-1478-1ada-7bde-928fd7cecc77, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.579606] env[62753]: DEBUG nova.network.neutron [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Updated VIF entry in instance network info cache for port f07bc1fd-c64f-4525-b198-0d7cc81d6927. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 882.579606] env[62753]: DEBUG nova.network.neutron [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Updating instance_info_cache with network_info: [{"id": "2b4dd6f0-5b5b-47ee-b7cb-ecafe1190758", "address": "fa:16:3e:8f:a0:ec", "network": {"id": "ebb4962b-34ac-499f-afea-8131f81571e1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-220454372", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.149", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b01e80a17e843a98ee662157230ae17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2b4dd6f0-5b", "ovs_interfaceid": "2b4dd6f0-5b5b-47ee-b7cb-ecafe1190758", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "51f3d79a-315b-418b-ac51-a3b5f5c328f3", "address": "fa:16:3e:2c:06:61", "network": {"id": "3e5a3d6d-6c74-4d41-91b4-136e08b33e52", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-1495343464", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4b01e80a17e843a98ee662157230ae17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "3d2e4070-a78e-4d08-a104-b6312ab65577", "external-id": "nsx-vlan-transportzone-292", "segmentation_id": 292, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap51f3d79a-31", "ovs_interfaceid": "51f3d79a-315b-418b-ac51-a3b5f5c328f3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "f07bc1fd-c64f-4525-b198-0d7cc81d6927", "address": "fa:16:3e:13:eb:0e", "network": {"id": "ebb4962b-34ac-499f-afea-8131f81571e1", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-220454372", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.207", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b01e80a17e843a98ee662157230ae17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c24464bb-bb6b-43a2-bdcd-8086ad1a307f", "external-id": "nsx-vlan-transportzone-781", "segmentation_id": 781, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf07bc1fd-c6", "ovs_interfaceid": "f07bc1fd-c64f-4525-b198-0d7cc81d6927", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 882.622151] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332251, 'name': CreateVM_Task, 'duration_secs': 0.370875} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 882.622628] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 882.623969] env[62753]: DEBUG oslo_concurrency.lockutils [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 882.749545] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332254, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 882.846507] env[62753]: DEBUG nova.compute.utils [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 882.852147] env[62753]: DEBUG nova.compute.manager [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Not allocating networking since 'none' was specified. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1977}} [ 882.924585] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332245, 'name': RelocateVM_Task} progress is 69%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.009618] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 883.032796] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a16416-1478-1ada-7bde-928fd7cecc77, 'name': SearchDatastore_Task, 'duration_secs': 0.012633} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.033189] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.033452] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 883.033703] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.033857] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.034083] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 883.034401] env[62753]: DEBUG oslo_concurrency.lockutils [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.034754] env[62753]: DEBUG oslo_concurrency.lockutils [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 883.035018] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2b1a39be-42a1-4f2c-9017-f18734d0e1dc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.037541] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-88d09ffb-46a4-4abc-9b2a-e3a9ae83fa3d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.046477] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 883.046477] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a03c65-721f-bd57-f701-c74ecc67f676" [ 883.046477] env[62753]: _type = "Task" [ 883.046477] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.047822] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 883.048025] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 883.052222] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6962ac83-d466-4147-bfac-bdd4e9871af5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.061296] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a03c65-721f-bd57-f701-c74ecc67f676, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.062813] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 883.062813] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5242f617-fa69-5cce-ed61-5e41ecdf1f65" [ 883.062813] env[62753]: _type = "Task" [ 883.062813] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.073330] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5242f617-fa69-5cce-ed61-5e41ecdf1f65, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.082297] env[62753]: DEBUG oslo_concurrency.lockutils [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] Releasing lock "refresh_cache-49311ef5-733a-4146-87ac-876a121a8d4d" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.082665] env[62753]: DEBUG nova.compute.manager [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Received event network-vif-plugged-9c52b1e4-82d2-4449-91f3-a624596bdc79 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.082977] env[62753]: DEBUG oslo_concurrency.lockutils [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] Acquiring lock "4d20e083-2959-453a-8875-47955bc02613-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.083227] env[62753]: DEBUG oslo_concurrency.lockutils [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] Lock "4d20e083-2959-453a-8875-47955bc02613-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.083401] env[62753]: DEBUG oslo_concurrency.lockutils [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] Lock "4d20e083-2959-453a-8875-47955bc02613-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.083581] env[62753]: DEBUG nova.compute.manager [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] No waiting events found dispatching network-vif-plugged-9c52b1e4-82d2-4449-91f3-a624596bdc79 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 883.083801] env[62753]: WARNING nova.compute.manager [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Received unexpected event network-vif-plugged-9c52b1e4-82d2-4449-91f3-a624596bdc79 for instance with vm_state building and task_state spawning. [ 883.084007] env[62753]: DEBUG nova.compute.manager [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Received event network-changed-9c52b1e4-82d2-4449-91f3-a624596bdc79 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 883.084244] env[62753]: DEBUG nova.compute.manager [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Refreshing instance network info cache due to event network-changed-9c52b1e4-82d2-4449-91f3-a624596bdc79. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 883.084466] env[62753]: DEBUG oslo_concurrency.lockutils [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] Acquiring lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.084616] env[62753]: DEBUG oslo_concurrency.lockutils [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] Acquired lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.084804] env[62753]: DEBUG nova.network.neutron [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Refreshing network info cache for port 9c52b1e4-82d2-4449-91f3-a624596bdc79 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 883.095446] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "a6ee010c-6305-4009-80e0-92a2c58bcd7b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.095703] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "a6ee010c-6305-4009-80e0-92a2c58bcd7b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.095936] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "a6ee010c-6305-4009-80e0-92a2c58bcd7b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.096160] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "a6ee010c-6305-4009-80e0-92a2c58bcd7b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.096354] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "a6ee010c-6305-4009-80e0-92a2c58bcd7b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 883.098570] env[62753]: INFO nova.compute.manager [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Terminating instance [ 883.100689] env[62753]: DEBUG nova.compute.manager [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 883.100898] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 883.102234] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3578fddd-07d3-4cbe-bdb2-4f697eda72e2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.116014] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 883.116349] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5ff83ec9-5bfb-4718-825b-4c4941190bc7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.126365] env[62753]: DEBUG oslo_vmware.api [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 883.126365] env[62753]: value = "task-1332255" [ 883.126365] env[62753]: _type = "Task" [ 883.126365] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.139419] env[62753]: DEBUG oslo_vmware.api [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332255, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.251992] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332254, 'name': CreateVM_Task, 'duration_secs': 0.930256} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.252199] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 883.253166] env[62753]: DEBUG oslo_concurrency.lockutils [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.253351] env[62753]: DEBUG oslo_concurrency.lockutils [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.253678] env[62753]: DEBUG oslo_concurrency.lockutils [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 883.253988] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c578158-88e4-4a26-ae0d-1a87cfceff76 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.263795] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 883.263795] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5232a214-9299-a18d-4454-eade327e5085" [ 883.263795] env[62753]: _type = "Task" [ 883.263795] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.277344] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5232a214-9299-a18d-4454-eade327e5085, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.351718] env[62753]: DEBUG nova.compute.manager [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 883.430412] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332245, 'name': RelocateVM_Task} progress is 84%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.567239] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a03c65-721f-bd57-f701-c74ecc67f676, 'name': SearchDatastore_Task, 'duration_secs': 0.017088} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.573361] env[62753]: DEBUG oslo_concurrency.lockutils [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.573619] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 883.573850] env[62753]: DEBUG oslo_concurrency.lockutils [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.583601] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5242f617-fa69-5cce-ed61-5e41ecdf1f65, 'name': SearchDatastore_Task, 'duration_secs': 0.015658} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.589206] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4bc8942a-0e0a-4ab1-83c0-b4d55f82bf07 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.600177] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 883.600177] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]527b10b4-ce39-2685-73c7-9b6aaef0759c" [ 883.600177] env[62753]: _type = "Task" [ 883.600177] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.611832] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]527b10b4-ce39-2685-73c7-9b6aaef0759c, 'name': SearchDatastore_Task, 'duration_secs': 0.010867} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.614778] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.615089] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 4d20e083-2959-453a-8875-47955bc02613/4d20e083-2959-453a-8875-47955bc02613.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 883.616572] env[62753]: DEBUG oslo_concurrency.lockutils [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.616572] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 883.616572] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a201f6df-d421-4d74-9853-44e746e00229 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.618097] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f4dfae7c-8a94-494c-af8f-9c46b64addd4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.631746] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 883.632037] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 883.633806] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-229a113d-f3b0-479b-b44e-07ef23325122 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.636578] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 883.636578] env[62753]: value = "task-1332256" [ 883.636578] env[62753]: _type = "Task" [ 883.636578] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.640066] env[62753]: DEBUG oslo_vmware.api [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332255, 'name': PowerOffVM_Task, 'duration_secs': 0.247191} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.646113] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 883.646358] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 883.646897] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8c5b7fb3-7741-423d-90a9-0b6deb4b9320 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.650029] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 883.650029] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52da093e-f8eb-0510-9bc3-6dd172db0799" [ 883.650029] env[62753]: _type = "Task" [ 883.650029] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.656890] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332256, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.658800] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7790f3b-533a-4677-a717-1625baf4766b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.667674] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52da093e-f8eb-0510-9bc3-6dd172db0799, 'name': SearchDatastore_Task, 'duration_secs': 0.009757} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.668398] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-10c16452-bae7-499f-821e-e0f163a28ad7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.674184] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c74a5dd-2c9d-4897-bf23-0566b868ceb5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.678952] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 883.678952] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ef20cd-ede1-53cf-f174-aaaefb74aea2" [ 883.678952] env[62753]: _type = "Task" [ 883.678952] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.709879] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3d4056b-4026-49f0-9cb3-397c14c56866 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.720904] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ef20cd-ede1-53cf-f174-aaaefb74aea2, 'name': SearchDatastore_Task, 'duration_secs': 0.010261} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.723800] env[62753]: DEBUG oslo_concurrency.lockutils [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.724092] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 50e1c1a1-7b8f-49cd-932d-03c920209634/50e1c1a1-7b8f-49cd-932d-03c920209634.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 883.724435] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 883.724620] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 883.724791] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Deleting the datastore file [datastore1] a6ee010c-6305-4009-80e0-92a2c58bcd7b {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 883.725039] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-32e0a5fd-6323-4e3f-8889-7a2b08a46d79 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.727981] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f1bbb47-4e69-4343-aba4-fad25b5f83cb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.731888] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-04fc8e47-bb00-4cce-b1d8-0d39a0b12d70 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.743851] env[62753]: DEBUG nova.compute.provider_tree [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 883.747739] env[62753]: DEBUG oslo_vmware.api [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 883.747739] env[62753]: value = "task-1332258" [ 883.747739] env[62753]: _type = "Task" [ 883.747739] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.748026] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 883.748026] env[62753]: value = "task-1332259" [ 883.748026] env[62753]: _type = "Task" [ 883.748026] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.763767] env[62753]: DEBUG oslo_vmware.api [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332258, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.766903] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332259, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.775595] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5232a214-9299-a18d-4454-eade327e5085, 'name': SearchDatastore_Task, 'duration_secs': 0.025312} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.775927] env[62753]: DEBUG oslo_concurrency.lockutils [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 883.776199] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 883.776479] env[62753]: DEBUG oslo_concurrency.lockutils [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 883.776622] env[62753]: DEBUG oslo_concurrency.lockutils [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 883.776798] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 883.777099] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3095d5e8-3cd8-41cc-86e1-5957c184f784 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.784750] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 883.784966] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 883.785770] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da84f738-2519-4e8c-93eb-bae4383fe8a9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.790687] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 883.790687] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]527a2038-89fc-8da8-ed4a-70ae5ed00734" [ 883.790687] env[62753]: _type = "Task" [ 883.790687] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.802434] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]527a2038-89fc-8da8-ed4a-70ae5ed00734, 'name': SearchDatastore_Task, 'duration_secs': 0.00891} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 883.803353] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b979520a-507c-4fdf-b1ef-e5f2ddfe5e58 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 883.808351] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 883.808351] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]521dfb09-c2f1-4c36-64ba-bdf214a3918c" [ 883.808351] env[62753]: _type = "Task" [ 883.808351] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 883.817461] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]521dfb09-c2f1-4c36-64ba-bdf214a3918c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 883.866626] env[62753]: DEBUG oslo_concurrency.lockutils [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "a401909c-5978-492b-a4fa-b17d38a5af44" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 883.866856] env[62753]: DEBUG oslo_concurrency.lockutils [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "a401909c-5978-492b-a4fa-b17d38a5af44" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 883.875106] env[62753]: DEBUG nova.network.neutron [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Updated VIF entry in instance network info cache for port 9c52b1e4-82d2-4449-91f3-a624596bdc79. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 883.875447] env[62753]: DEBUG nova.network.neutron [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Updating instance_info_cache with network_info: [{"id": "9c52b1e4-82d2-4449-91f3-a624596bdc79", "address": "fa:16:3e:8e:1b:de", "network": {"id": "12446a5e-7329-4f8f-aa28-a8b9d78e7573", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1766107072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "371a5e8b16be4cffbdeb2139757684fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7c7f16b-a2e3-4d1f-9b7e-a44a9ebb589c", "external-id": "nsx-vlan-transportzone-772", "segmentation_id": 772, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c52b1e4-82", "ovs_interfaceid": "9c52b1e4-82d2-4449-91f3-a624596bdc79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 883.923995] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332245, 'name': RelocateVM_Task} progress is 97%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.010029] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 884.010029] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 884.010029] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 884.010029] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62753) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 884.149694] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332256, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.249646] env[62753]: DEBUG nova.scheduler.client.report [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 884.265025] env[62753]: DEBUG oslo_vmware.api [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332258, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.177229} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.268418] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 884.268663] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 884.268876] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 884.269093] env[62753]: INFO nova.compute.manager [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Took 1.17 seconds to destroy the instance on the hypervisor. [ 884.269395] env[62753]: DEBUG oslo.service.loopingcall [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.269620] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332259, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.270115] env[62753]: DEBUG nova.compute.manager [-] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 884.270219] env[62753]: DEBUG nova.network.neutron [-] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 884.320184] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]521dfb09-c2f1-4c36-64ba-bdf214a3918c, 'name': SearchDatastore_Task, 'duration_secs': 0.010176} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.320494] env[62753]: DEBUG oslo_concurrency.lockutils [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.320837] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 49311ef5-733a-4146-87ac-876a121a8d4d/49311ef5-733a-4146-87ac-876a121a8d4d.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 884.321047] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-94cfe2ca-af20-40fd-94ad-0e3b8a76bd80 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.330559] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 884.330559] env[62753]: value = "task-1332260" [ 884.330559] env[62753]: _type = "Task" [ 884.330559] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.343147] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332260, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.368652] env[62753]: DEBUG nova.compute.manager [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 884.378357] env[62753]: DEBUG oslo_concurrency.lockutils [req-77e11735-4bc2-4623-b638-3eca8d36b27e req-f431bac3-6de9-4b23-a196-b0192ca6be90 service nova] Releasing lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 884.404141] env[62753]: DEBUG nova.virt.hardware [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 884.404779] env[62753]: DEBUG nova.virt.hardware [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 884.405046] env[62753]: DEBUG nova.virt.hardware [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 884.405323] env[62753]: DEBUG nova.virt.hardware [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 884.405524] env[62753]: DEBUG nova.virt.hardware [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 884.405799] env[62753]: DEBUG nova.virt.hardware [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 884.406138] env[62753]: DEBUG nova.virt.hardware [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 884.406410] env[62753]: DEBUG nova.virt.hardware [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 884.406695] env[62753]: DEBUG nova.virt.hardware [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 884.406997] env[62753]: DEBUG nova.virt.hardware [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 884.407301] env[62753]: DEBUG nova.virt.hardware [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 884.409012] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f3f12c5-9b84-440d-a5f2-ffa71135e095 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.422615] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-16edd898-45c8-4124-bea3-cb2dfb3a374c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.437879] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Instance VIF info [] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 884.444679] env[62753]: DEBUG oslo.service.loopingcall [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 884.448981] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 884.449436] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332245, 'name': RelocateVM_Task} progress is 98%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.449703] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-26a5a96f-b7b3-47e3-93ba-f601bf242a10 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.469437] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 884.469437] env[62753]: value = "task-1332261" [ 884.469437] env[62753]: _type = "Task" [ 884.469437] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.480295] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332261, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.653727] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332256, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.761069] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.422s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 884.761755] env[62753]: DEBUG nova.compute.manager [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 884.769579] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.892s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 884.773721] env[62753]: INFO nova.compute.claims [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 884.778424] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332259, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.016577} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.779087] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 50e1c1a1-7b8f-49cd-932d-03c920209634/50e1c1a1-7b8f-49cd-932d-03c920209634.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 884.779318] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 884.779792] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f411aca-7b14-43a0-9ee9-3e08c27dbf2f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.788246] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 884.788246] env[62753]: value = "task-1332262" [ 884.788246] env[62753]: _type = "Task" [ 884.788246] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 884.797979] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332262, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.831299] env[62753]: DEBUG nova.compute.manager [req-7575fdfa-75b3-480c-9df1-d8dbe25482ab req-12295ad6-9d43-4d99-88c3-755d4fcecfac service nova] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Received event network-vif-deleted-bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 884.832669] env[62753]: INFO nova.compute.manager [req-7575fdfa-75b3-480c-9df1-d8dbe25482ab req-12295ad6-9d43-4d99-88c3-755d4fcecfac service nova] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Neutron deleted interface bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c; detaching it from the instance and deleting it from the info cache [ 884.833627] env[62753]: DEBUG nova.network.neutron [req-7575fdfa-75b3-480c-9df1-d8dbe25482ab req-12295ad6-9d43-4d99-88c3-755d4fcecfac service nova] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 884.850843] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332260, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 884.926787] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332245, 'name': RelocateVM_Task, 'duration_secs': 3.123816} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.927124] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Volume attach. Driver type: vmdk {{(pid=62753) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 884.927343] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284559', 'volume_id': '02056526-1ec1-42b1-96d8-e237f65687ce', 'name': 'volume-02056526-1ec1-42b1-96d8-e237f65687ce', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c5dae9a3-052b-4f4c-ac79-84aeaded457a', 'attached_at': '', 'detached_at': '', 'volume_id': '02056526-1ec1-42b1-96d8-e237f65687ce', 'serial': '02056526-1ec1-42b1-96d8-e237f65687ce'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 884.928190] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-917749ef-7c55-42dd-8496-fcec02a68da1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.946422] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-51e5f918-da90-4651-9bb4-ea8b7a72dce8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.968942] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Reconfiguring VM instance instance-00000040 to attach disk [datastore2] volume-02056526-1ec1-42b1-96d8-e237f65687ce/volume-02056526-1ec1-42b1-96d8-e237f65687ce.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 884.969405] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-73647fb7-7795-43bf-bc0c-fceffbc2d0a3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 884.998120] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332261, 'name': CreateVM_Task, 'duration_secs': 0.439699} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 884.999496] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 884.999905] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Waiting for the task: (returnval){ [ 884.999905] env[62753]: value = "task-1332263" [ 884.999905] env[62753]: _type = "Task" [ 884.999905] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.000350] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.000533] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.000869] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 885.001239] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b0ceb07-c5ea-4d88-9153-1968afdfee88 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.006281] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 885.009730] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 885.011636] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 885.013225] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 885.013225] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52947353-1845-7d00-0360-6c64c584c270" [ 885.013225] env[62753]: _type = "Task" [ 885.013225] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.017330] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332263, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.028832] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52947353-1845-7d00-0360-6c64c584c270, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.150204] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332256, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.131879} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.150460] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 4d20e083-2959-453a-8875-47955bc02613/4d20e083-2959-453a-8875-47955bc02613.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 885.150690] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 885.150962] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b475b48c-ae01-4fae-a60d-fe3bfc231284 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.157098] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 885.157098] env[62753]: value = "task-1332264" [ 885.157098] env[62753]: _type = "Task" [ 885.157098] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.165229] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332264, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.279905] env[62753]: DEBUG nova.network.neutron [-] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 885.282283] env[62753]: DEBUG nova.compute.utils [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 885.285411] env[62753]: DEBUG nova.compute.manager [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 885.285575] env[62753]: DEBUG nova.network.neutron [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 885.300331] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332262, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076045} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.300694] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 885.301582] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c95b1fd-df7b-4d84-8a68-2cfd149c8368 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.321850] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Reconfiguring VM instance instance-00000042 to attach disk [datastore2] 50e1c1a1-7b8f-49cd-932d-03c920209634/50e1c1a1-7b8f-49cd-932d-03c920209634.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 885.322169] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-29d1392a-2182-4639-b038-18e3ad746a97 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.339445] env[62753]: DEBUG nova.policy [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd88543296c8c4ad4be4f3e3ea34f7779', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'f0fb90849a6c494996e4fba4ea404cc4', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 885.348936] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-bca0cff3-41a3-425c-af37-a645ca6f860c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.349884] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 885.349884] env[62753]: value = "task-1332265" [ 885.349884] env[62753]: _type = "Task" [ 885.349884] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.354201] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332260, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.775315} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.357603] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 49311ef5-733a-4146-87ac-876a121a8d4d/49311ef5-733a-4146-87ac-876a121a8d4d.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 885.357908] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 885.359188] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-68dd2a1a-78af-4793-9db0-3c634f192a8e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.365027] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932532c6-fb39-411e-b14c-c0224022edeb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.383327] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332265, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.387177] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 885.387177] env[62753]: value = "task-1332266" [ 885.387177] env[62753]: _type = "Task" [ 885.387177] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.395758] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332266, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.403266] env[62753]: DEBUG nova.compute.manager [req-7575fdfa-75b3-480c-9df1-d8dbe25482ab req-12295ad6-9d43-4d99-88c3-755d4fcecfac service nova] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Detach interface failed, port_id=bcc79f72-e1a6-426e-9e71-e1aaab3bfe2c, reason: Instance a6ee010c-6305-4009-80e0-92a2c58bcd7b could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 885.511948] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332263, 'name': ReconfigVM_Task, 'duration_secs': 0.320294} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.512270] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Reconfigured VM instance instance-00000040 to attach disk [datastore2] volume-02056526-1ec1-42b1-96d8-e237f65687ce/volume-02056526-1ec1-42b1-96d8-e237f65687ce.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 885.517794] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 885.518183] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5fb35eb7-7c81-433b-8e73-aecd7f8b8cfa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.537350] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52947353-1845-7d00-0360-6c64c584c270, 'name': SearchDatastore_Task, 'duration_secs': 0.019933} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.538630] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 885.538982] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 885.539388] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 885.539705] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 885.539925] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 885.540304] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Waiting for the task: (returnval){ [ 885.540304] env[62753]: value = "task-1332267" [ 885.540304] env[62753]: _type = "Task" [ 885.540304] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.540598] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-800ed6ba-45c2-4bd8-90fe-96af23250fad {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.552062] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332267, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.553263] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 885.553440] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 885.557142] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-836777f8-ac91-4f08-93de-432c0ccf2f63 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.559257] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 885.559257] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]521a1de1-81d2-8018-8c93-18f42b3ed7ef" [ 885.559257] env[62753]: _type = "Task" [ 885.559257] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.567840] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]521a1de1-81d2-8018-8c93-18f42b3ed7ef, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.666451] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332264, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06124} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.666762] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 885.671701] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3df42104-56e5-4c57-ac7a-471c5fee3693 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.695218] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 4d20e083-2959-453a-8875-47955bc02613/4d20e083-2959-453a-8875-47955bc02613.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 885.695514] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-535d6023-56c5-4aeb-b4e4-7dd9efecb739 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.714488] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 885.714488] env[62753]: value = "task-1332268" [ 885.714488] env[62753]: _type = "Task" [ 885.714488] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.722958] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332268, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.784734] env[62753]: INFO nova.compute.manager [-] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Took 1.51 seconds to deallocate network for instance. [ 885.790924] env[62753]: DEBUG nova.compute.manager [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 885.800292] env[62753]: DEBUG nova.network.neutron [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Successfully created port: 1f18d3a5-8537-41a7-adcc-ea26fe748a8e {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 885.866154] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332265, 'name': ReconfigVM_Task, 'duration_secs': 0.272187} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.869355] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Reconfigured VM instance instance-00000042 to attach disk [datastore2] 50e1c1a1-7b8f-49cd-932d-03c920209634/50e1c1a1-7b8f-49cd-932d-03c920209634.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 885.870145] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-98c595b8-2dc4-4332-a5bc-bef70f9d177a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.876212] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 885.876212] env[62753]: value = "task-1332269" [ 885.876212] env[62753]: _type = "Task" [ 885.876212] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.886307] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332269, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 885.897500] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332266, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.163} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 885.897894] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 885.898766] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c519c0d7-ef29-43d8-a5d4-b28c32c3dd87 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.927249] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Reconfiguring VM instance instance-0000003f to attach disk [datastore1] 49311ef5-733a-4146-87ac-876a121a8d4d/49311ef5-733a-4146-87ac-876a121a8d4d.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 885.930611] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a7b763d-2475-4890-ab4f-19273d2adef7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 885.951952] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 885.951952] env[62753]: value = "task-1332270" [ 885.951952] env[62753]: _type = "Task" [ 885.951952] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 885.959914] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332270, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.056197] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332267, 'name': ReconfigVM_Task, 'duration_secs': 0.128572} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.056525] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284559', 'volume_id': '02056526-1ec1-42b1-96d8-e237f65687ce', 'name': 'volume-02056526-1ec1-42b1-96d8-e237f65687ce', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c5dae9a3-052b-4f4c-ac79-84aeaded457a', 'attached_at': '', 'detached_at': '', 'volume_id': '02056526-1ec1-42b1-96d8-e237f65687ce', 'serial': '02056526-1ec1-42b1-96d8-e237f65687ce'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 886.057113] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2376527b-0028-4567-bc87-44e4ba88fdf0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.066048] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Waiting for the task: (returnval){ [ 886.066048] env[62753]: value = "task-1332271" [ 886.066048] env[62753]: _type = "Task" [ 886.066048] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.073927] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]521a1de1-81d2-8018-8c93-18f42b3ed7ef, 'name': SearchDatastore_Task, 'duration_secs': 0.008873} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.077560] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332271, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.077724] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c8b5540-ef0a-4ef2-aefe-c74617997709 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.084969] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 886.084969] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523dd6a5-080f-1caa-583b-25a078bee6bd" [ 886.084969] env[62753]: _type = "Task" [ 886.084969] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.092643] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523dd6a5-080f-1caa-583b-25a078bee6bd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.119579] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-241573c6-012c-4ede-984b-37cdd2038b3a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.128702] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab0b7233-2e0c-48b6-804a-3e77b2714e43 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.161230] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a16d49a0-9a6e-4ddc-8e43-c93e5e00a4d6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.169022] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1df8a93c-0cbc-4915-946c-935e66ed6ff3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.182538] env[62753]: DEBUG nova.compute.provider_tree [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 886.223486] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332268, 'name': ReconfigVM_Task, 'duration_secs': 0.290752} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.223766] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 4d20e083-2959-453a-8875-47955bc02613/4d20e083-2959-453a-8875-47955bc02613.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.224651] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c17a9929-21a8-4e5f-beca-d0d9859dd05a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.230923] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 886.230923] env[62753]: value = "task-1332272" [ 886.230923] env[62753]: _type = "Task" [ 886.230923] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.238671] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332272, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.306605] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 886.387028] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332269, 'name': Rename_Task, 'duration_secs': 0.183823} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.387141] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 886.387315] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-76bcf2e2-3947-4eff-a104-06f62e936c2a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.394124] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 886.394124] env[62753]: value = "task-1332273" [ 886.394124] env[62753]: _type = "Task" [ 886.394124] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.401741] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332273, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.461186] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332270, 'name': ReconfigVM_Task, 'duration_secs': 0.309981} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.461503] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Reconfigured VM instance instance-0000003f to attach disk [datastore1] 49311ef5-733a-4146-87ac-876a121a8d4d/49311ef5-733a-4146-87ac-876a121a8d4d.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 886.462191] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-58f55c6a-835d-438e-ac13-97b7a0c68076 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.469120] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 886.469120] env[62753]: value = "task-1332274" [ 886.469120] env[62753]: _type = "Task" [ 886.469120] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.476708] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332274, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.576104] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332271, 'name': Rename_Task, 'duration_secs': 0.131039} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.576379] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 886.576628] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-59b9f880-75b0-4126-970e-879f89fda2f8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.583881] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Waiting for the task: (returnval){ [ 886.583881] env[62753]: value = "task-1332275" [ 886.583881] env[62753]: _type = "Task" [ 886.583881] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.594876] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332275, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.598911] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523dd6a5-080f-1caa-583b-25a078bee6bd, 'name': SearchDatastore_Task, 'duration_secs': 0.009262} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.599191] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 886.599454] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 33c1de85-aade-4c69-8fdb-7672d8c53300/33c1de85-aade-4c69-8fdb-7672d8c53300.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 886.599719] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-11a5fcd6-e80b-46f9-a2fd-2c8d07554ac8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.606299] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 886.606299] env[62753]: value = "task-1332276" [ 886.606299] env[62753]: _type = "Task" [ 886.606299] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.614476] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332276, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.686014] env[62753]: DEBUG nova.scheduler.client.report [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 886.743054] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332272, 'name': Rename_Task, 'duration_secs': 0.147944} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.743379] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 886.743657] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a80b5716-69da-4c85-b4ec-b96f58b37579 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.749749] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 886.749749] env[62753]: value = "task-1332277" [ 886.749749] env[62753]: _type = "Task" [ 886.749749] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 886.758287] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332277, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 886.811081] env[62753]: DEBUG nova.compute.manager [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 886.836935] env[62753]: DEBUG nova.virt.hardware [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 886.837190] env[62753]: DEBUG nova.virt.hardware [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 886.837359] env[62753]: DEBUG nova.virt.hardware [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 886.837553] env[62753]: DEBUG nova.virt.hardware [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 886.837704] env[62753]: DEBUG nova.virt.hardware [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 886.837860] env[62753]: DEBUG nova.virt.hardware [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 886.838096] env[62753]: DEBUG nova.virt.hardware [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 886.838269] env[62753]: DEBUG nova.virt.hardware [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 886.838630] env[62753]: DEBUG nova.virt.hardware [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 886.838863] env[62753]: DEBUG nova.virt.hardware [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 886.839092] env[62753]: DEBUG nova.virt.hardware [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 886.840403] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214af6fb-0515-489d-b7fd-4e1dc54b8dde {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.851669] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-529e3ab4-bb14-47db-8f4f-e280e4486cae {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.905796] env[62753]: DEBUG oslo_vmware.api [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332273, 'name': PowerOnVM_Task, 'duration_secs': 0.484832} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.906106] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 886.906343] env[62753]: INFO nova.compute.manager [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Took 4.91 seconds to spawn the instance on the hypervisor. [ 886.906575] env[62753]: DEBUG nova.compute.manager [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 886.907342] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b34117d2-a619-457d-9676-5b53ab0dd20c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.983854] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332274, 'name': Rename_Task, 'duration_secs': 0.14777} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 886.984749] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 886.985126] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6c0b260b-1653-463a-87b1-7a39356b1694 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 886.995421] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 886.995421] env[62753]: value = "task-1332278" [ 886.995421] env[62753]: _type = "Task" [ 886.995421] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.008423] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332278, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.097208] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332275, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.118694] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332276, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.509193} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.118986] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 33c1de85-aade-4c69-8fdb-7672d8c53300/33c1de85-aade-4c69-8fdb-7672d8c53300.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 887.119271] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 887.119486] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3347f60c-1914-4fc7-bcfe-8d03777df7d3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.126452] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 887.126452] env[62753]: value = "task-1332279" [ 887.126452] env[62753]: _type = "Task" [ 887.126452] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.134286] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332279, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.191605] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.422s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.192194] env[62753]: DEBUG nova.compute.manager [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 887.195109] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.306s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.196707] env[62753]: INFO nova.compute.claims [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 887.260948] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332277, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.391034] env[62753]: DEBUG nova.compute.manager [req-caf5bfc2-ea02-4f62-99dc-12625c0c8cbc req-18b3786c-a578-4eea-8b81-43ac581912b0 service nova] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Received event network-vif-plugged-1f18d3a5-8537-41a7-adcc-ea26fe748a8e {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 887.391271] env[62753]: DEBUG oslo_concurrency.lockutils [req-caf5bfc2-ea02-4f62-99dc-12625c0c8cbc req-18b3786c-a578-4eea-8b81-43ac581912b0 service nova] Acquiring lock "57eef548-a0f1-4f0b-a026-885a10d005c2-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 887.391487] env[62753]: DEBUG oslo_concurrency.lockutils [req-caf5bfc2-ea02-4f62-99dc-12625c0c8cbc req-18b3786c-a578-4eea-8b81-43ac581912b0 service nova] Lock "57eef548-a0f1-4f0b-a026-885a10d005c2-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 887.391663] env[62753]: DEBUG oslo_concurrency.lockutils [req-caf5bfc2-ea02-4f62-99dc-12625c0c8cbc req-18b3786c-a578-4eea-8b81-43ac581912b0 service nova] Lock "57eef548-a0f1-4f0b-a026-885a10d005c2-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.391835] env[62753]: DEBUG nova.compute.manager [req-caf5bfc2-ea02-4f62-99dc-12625c0c8cbc req-18b3786c-a578-4eea-8b81-43ac581912b0 service nova] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] No waiting events found dispatching network-vif-plugged-1f18d3a5-8537-41a7-adcc-ea26fe748a8e {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 887.392010] env[62753]: WARNING nova.compute.manager [req-caf5bfc2-ea02-4f62-99dc-12625c0c8cbc req-18b3786c-a578-4eea-8b81-43ac581912b0 service nova] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Received unexpected event network-vif-plugged-1f18d3a5-8537-41a7-adcc-ea26fe748a8e for instance with vm_state building and task_state spawning. [ 887.425559] env[62753]: INFO nova.compute.manager [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Took 31.67 seconds to build instance. [ 887.492730] env[62753]: DEBUG nova.network.neutron [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Successfully updated port: 1f18d3a5-8537-41a7-adcc-ea26fe748a8e {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 887.505139] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332278, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.594729] env[62753]: DEBUG oslo_vmware.api [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332275, 'name': PowerOnVM_Task, 'duration_secs': 0.656049} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.595040] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 887.595260] env[62753]: INFO nova.compute.manager [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Took 8.63 seconds to spawn the instance on the hypervisor. [ 887.595445] env[62753]: DEBUG nova.compute.manager [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 887.596224] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b32670fd-42f3-48bb-bfa6-c81f5e8adf03 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.636813] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332279, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071084} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 887.637126] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 887.637896] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47838589-dd1c-4fe4-aacb-8933a65daaa7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.658370] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 33c1de85-aade-4c69-8fdb-7672d8c53300/33c1de85-aade-4c69-8fdb-7672d8c53300.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 887.659162] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-80e49052-8023-4f7e-b3a3-386e199ad815 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 887.680069] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 887.680069] env[62753]: value = "task-1332280" [ 887.680069] env[62753]: _type = "Task" [ 887.680069] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 887.688161] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332280, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.700918] env[62753]: DEBUG nova.compute.utils [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 887.704381] env[62753]: DEBUG nova.compute.manager [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 887.704565] env[62753]: DEBUG nova.network.neutron [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 887.744911] env[62753]: DEBUG nova.policy [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae52823f7c93454e8089261c3ca44321', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d52d6eaee934be5ab0e0003df1ce316', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 887.760545] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332277, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 887.927541] env[62753]: DEBUG oslo_concurrency.lockutils [None req-644406b0-01f7-4f25-acc3-7b39a71993be tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "50e1c1a1-7b8f-49cd-932d-03c920209634" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.178s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 887.995873] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Acquiring lock "refresh_cache-57eef548-a0f1-4f0b-a026-885a10d005c2" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 887.996185] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Acquired lock "refresh_cache-57eef548-a0f1-4f0b-a026-885a10d005c2" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 887.996453] env[62753]: DEBUG nova.network.neutron [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 888.008803] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332278, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.117950] env[62753]: INFO nova.compute.manager [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Took 36.47 seconds to build instance. [ 888.190907] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332280, 'name': ReconfigVM_Task, 'duration_secs': 0.302205} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.193704] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 33c1de85-aade-4c69-8fdb-7672d8c53300/33c1de85-aade-4c69-8fdb-7672d8c53300.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 888.193704] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-81ee4d63-cbd6-492f-9ac9-6b1dbdd91f27 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.202691] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 888.202691] env[62753]: value = "task-1332281" [ 888.202691] env[62753]: _type = "Task" [ 888.202691] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.205726] env[62753]: DEBUG nova.compute.manager [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 888.216475] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332281, 'name': Rename_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.256371] env[62753]: DEBUG nova.network.neutron [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Successfully created port: ce484d8b-65ec-4512-aaf1-48a0c51f4f74 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 888.268924] env[62753]: DEBUG oslo_vmware.api [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332277, 'name': PowerOnVM_Task, 'duration_secs': 1.165646} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.269225] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 888.269422] env[62753]: INFO nova.compute.manager [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Took 8.65 seconds to spawn the instance on the hypervisor. [ 888.269605] env[62753]: DEBUG nova.compute.manager [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 888.270411] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-76a8f462-3225-443b-b328-a98c7d95fd5c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.430928] env[62753]: DEBUG nova.compute.manager [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 888.498792] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46f31596-ebcc-4b42-b6f0-4721379c1cbd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.514481] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fadb6aa0-9285-477a-a18f-4e986f8fdfbf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.517889] env[62753]: DEBUG oslo_vmware.api [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332278, 'name': PowerOnVM_Task, 'duration_secs': 1.032203} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.518637] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 888.518637] env[62753]: INFO nova.compute.manager [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Took 13.75 seconds to spawn the instance on the hypervisor. [ 888.518637] env[62753]: DEBUG nova.compute.manager [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 888.519831] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-869dc972-d700-4579-adfe-3af561aec9bc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.548613] env[62753]: DEBUG nova.network.neutron [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 888.551352] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64845ea1-fc60-46d1-87b7-c0179516c5bf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.564025] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7587d8b3-3fa7-45e4-a9e1-1e810010f08a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.581281] env[62753]: DEBUG nova.compute.provider_tree [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 888.619777] env[62753]: DEBUG oslo_concurrency.lockutils [None req-258b583d-0d28-4487-a817-0ffb02ee7633 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Lock "c5dae9a3-052b-4f4c-ac79-84aeaded457a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 67.893s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 888.709215] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332281, 'name': Rename_Task, 'duration_secs': 0.132424} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 888.710096] env[62753]: DEBUG nova.network.neutron [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Updating instance_info_cache with network_info: [{"id": "1f18d3a5-8537-41a7-adcc-ea26fe748a8e", "address": "fa:16:3e:54:35:11", "network": {"id": "01ef394e-a94d-4d7e-bfe7-4d73f44c8320", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-383531650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0fb90849a6c494996e4fba4ea404cc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f18d3a5-85", "ovs_interfaceid": "1f18d3a5-8537-41a7-adcc-ea26fe748a8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 888.711285] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 888.711723] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-27682138-3e8a-4932-aed8-cb5d0a01bfac {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 888.724014] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 888.724014] env[62753]: value = "task-1332282" [ 888.724014] env[62753]: _type = "Task" [ 888.724014] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 888.729426] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332282, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 888.788274] env[62753]: INFO nova.compute.manager [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Took 34.88 seconds to build instance. [ 888.878111] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ef0d156-39c9-44ed-b120-38d98e69a692 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "interface-3bc05fdb-81be-4764-b9d9-01acd125b020-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 888.878443] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ef0d156-39c9-44ed-b120-38d98e69a692 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "interface-3bc05fdb-81be-4764-b9d9-01acd125b020-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 888.878848] env[62753]: DEBUG nova.objects.instance [None req-9ef0d156-39c9-44ed-b120-38d98e69a692 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lazy-loading 'flavor' on Instance uuid 3bc05fdb-81be-4764-b9d9-01acd125b020 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 888.956033] env[62753]: DEBUG oslo_concurrency.lockutils [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.068277] env[62753]: INFO nova.compute.manager [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Took 41.32 seconds to build instance. [ 889.084212] env[62753]: DEBUG nova.scheduler.client.report [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 889.123256] env[62753]: DEBUG nova.compute.manager [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 889.212938] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Releasing lock "refresh_cache-57eef548-a0f1-4f0b-a026-885a10d005c2" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 889.213300] env[62753]: DEBUG nova.compute.manager [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Instance network_info: |[{"id": "1f18d3a5-8537-41a7-adcc-ea26fe748a8e", "address": "fa:16:3e:54:35:11", "network": {"id": "01ef394e-a94d-4d7e-bfe7-4d73f44c8320", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-383531650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0fb90849a6c494996e4fba4ea404cc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f18d3a5-85", "ovs_interfaceid": "1f18d3a5-8537-41a7-adcc-ea26fe748a8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 889.213729] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:54:35:11', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c894ab55-c869-4530-9702-cb46d173ce94', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '1f18d3a5-8537-41a7-adcc-ea26fe748a8e', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 889.221210] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Creating folder: Project (f0fb90849a6c494996e4fba4ea404cc4). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 889.222700] env[62753]: DEBUG nova.compute.manager [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 889.225281] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-22ae3b4f-f0a0-4c18-82b7-9d62f1f33666 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.236980] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332282, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.238873] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Created folder: Project (f0fb90849a6c494996e4fba4ea404cc4) in parent group-v284541. [ 889.238873] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Creating folder: Instances. Parent ref: group-v284595. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 889.238873] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-c36621db-21d4-4bef-9114-761246bb8f83 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.249146] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Created folder: Instances in parent group-v284595. [ 889.249392] env[62753]: DEBUG oslo.service.loopingcall [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 889.249588] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 889.249970] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ae5180ac-e710-41c8-80b0-f9a3b2d9cac5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.272537] env[62753]: DEBUG nova.virt.hardware [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 889.272844] env[62753]: DEBUG nova.virt.hardware [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 889.273061] env[62753]: DEBUG nova.virt.hardware [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 889.273281] env[62753]: DEBUG nova.virt.hardware [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 889.273492] env[62753]: DEBUG nova.virt.hardware [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 889.273739] env[62753]: DEBUG nova.virt.hardware [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 889.274023] env[62753]: DEBUG nova.virt.hardware [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 889.274244] env[62753]: DEBUG nova.virt.hardware [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 889.274491] env[62753]: DEBUG nova.virt.hardware [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 889.274685] env[62753]: DEBUG nova.virt.hardware [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 889.275312] env[62753]: DEBUG nova.virt.hardware [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 889.275963] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9016c595-056d-4a38-86e4-14de0a6cd402 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.286668] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a0a3f74-67e2-48b9-bade-2583a8d3514c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.291038] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 889.291038] env[62753]: value = "task-1332285" [ 889.291038] env[62753]: _type = "Task" [ 889.291038] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.291230] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d53e2648-14c1-45f8-9957-31a2a499c207 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "4d20e083-2959-453a-8875-47955bc02613" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 61.247s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.311404] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332285, 'name': CreateVM_Task} progress is 10%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.384023] env[62753]: DEBUG nova.objects.instance [None req-9ef0d156-39c9-44ed-b120-38d98e69a692 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lazy-loading 'pci_requests' on Instance uuid 3bc05fdb-81be-4764-b9d9-01acd125b020 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 889.570626] env[62753]: DEBUG oslo_concurrency.lockutils [None req-08172429-56eb-4728-8401-38469ed06f6a tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "49311ef5-733a-4146-87ac-876a121a8d4d" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 90.885s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.589304] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.394s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 889.589891] env[62753]: DEBUG nova.compute.manager [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 889.592888] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 18.644s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 889.592888] env[62753]: DEBUG nova.objects.instance [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62753) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 889.640274] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 889.735546] env[62753]: DEBUG oslo_vmware.api [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332282, 'name': PowerOnVM_Task, 'duration_secs': 0.59756} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.736010] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 889.736278] env[62753]: INFO nova.compute.manager [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Took 5.37 seconds to spawn the instance on the hypervisor. [ 889.736467] env[62753]: DEBUG nova.compute.manager [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 889.737280] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec1bccdb-e84d-439d-b1c7-0f0ee0d25f1a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.796542] env[62753]: DEBUG nova.compute.manager [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 889.807027] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332285, 'name': CreateVM_Task, 'duration_secs': 0.32232} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 889.807027] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 889.807027] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.807027] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.807027] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 889.807027] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2f41603-ba04-4c14-8838-035233bafe0c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 889.811647] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Waiting for the task: (returnval){ [ 889.811647] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52b5019c-3020-bc06-a80b-c01a175d94a2" [ 889.811647] env[62753]: _type = "Task" [ 889.811647] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 889.820133] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52b5019c-3020-bc06-a80b-c01a175d94a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 889.885761] env[62753]: DEBUG nova.objects.base [None req-9ef0d156-39c9-44ed-b120-38d98e69a692 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Object Instance<3bc05fdb-81be-4764-b9d9-01acd125b020> lazy-loaded attributes: flavor,pci_requests {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 889.886016] env[62753]: DEBUG nova.network.neutron [None req-9ef0d156-39c9-44ed-b120-38d98e69a692 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 889.951638] env[62753]: DEBUG nova.compute.manager [req-983db690-6c81-4d86-bfd2-5547da2e7fb9 req-75b53db4-40ba-4447-912f-5c1e595dcc86 service nova] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Received event network-changed-1f18d3a5-8537-41a7-adcc-ea26fe748a8e {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 889.951848] env[62753]: DEBUG nova.compute.manager [req-983db690-6c81-4d86-bfd2-5547da2e7fb9 req-75b53db4-40ba-4447-912f-5c1e595dcc86 service nova] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Refreshing instance network info cache due to event network-changed-1f18d3a5-8537-41a7-adcc-ea26fe748a8e. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 889.952135] env[62753]: DEBUG oslo_concurrency.lockutils [req-983db690-6c81-4d86-bfd2-5547da2e7fb9 req-75b53db4-40ba-4447-912f-5c1e595dcc86 service nova] Acquiring lock "refresh_cache-57eef548-a0f1-4f0b-a026-885a10d005c2" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 889.952254] env[62753]: DEBUG oslo_concurrency.lockutils [req-983db690-6c81-4d86-bfd2-5547da2e7fb9 req-75b53db4-40ba-4447-912f-5c1e595dcc86 service nova] Acquired lock "refresh_cache-57eef548-a0f1-4f0b-a026-885a10d005c2" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 889.952369] env[62753]: DEBUG nova.network.neutron [req-983db690-6c81-4d86-bfd2-5547da2e7fb9 req-75b53db4-40ba-4447-912f-5c1e595dcc86 service nova] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Refreshing network info cache for port 1f18d3a5-8537-41a7-adcc-ea26fe748a8e {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 889.988844] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9ef0d156-39c9-44ed-b120-38d98e69a692 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "interface-3bc05fdb-81be-4764-b9d9-01acd125b020-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 1.110s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.073426] env[62753]: DEBUG nova.compute.manager [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 890.103013] env[62753]: DEBUG nova.compute.utils [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 890.105479] env[62753]: DEBUG nova.compute.manager [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 890.105748] env[62753]: DEBUG nova.network.neutron [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 890.151321] env[62753]: DEBUG nova.policy [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '0b3bb7961af443628176d8341f1fc2ca', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'db8f96fcee6e47e786ceae2ed3c0645b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 890.213478] env[62753]: DEBUG nova.network.neutron [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Successfully updated port: ce484d8b-65ec-4512-aaf1-48a0c51f4f74 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 890.254928] env[62753]: INFO nova.compute.manager [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Took 33.78 seconds to build instance. [ 890.321727] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52b5019c-3020-bc06-a80b-c01a175d94a2, 'name': SearchDatastore_Task, 'duration_secs': 0.010431} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.322018] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 890.322334] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 890.322501] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.322650] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.322885] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 890.323354] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8986c395-1c1e-425e-be12-1760cb5e687a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.326092] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.335245] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 890.335463] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 890.336216] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-61ecd203-f1a0-413f-b765-8caf9a0e7f99 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.341552] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Waiting for the task: (returnval){ [ 890.341552] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52475f92-57c9-425f-d599-8ca631a9c593" [ 890.341552] env[62753]: _type = "Task" [ 890.341552] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.349945] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52475f92-57c9-425f-d599-8ca631a9c593, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 890.414463] env[62753]: DEBUG nova.network.neutron [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Successfully created port: 61fe245f-aaa8-4566-857d-3002a5c3426d {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 890.592189] env[62753]: DEBUG oslo_concurrency.lockutils [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 890.609915] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0ef8d3f9-b66f-4032-9297-849f8538fce5 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.611362] env[62753]: DEBUG nova.compute.manager [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 890.614994] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.738s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 890.619024] env[62753]: INFO nova.compute.claims [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 890.716102] env[62753]: DEBUG nova.network.neutron [req-983db690-6c81-4d86-bfd2-5547da2e7fb9 req-75b53db4-40ba-4447-912f-5c1e595dcc86 service nova] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Updated VIF entry in instance network info cache for port 1f18d3a5-8537-41a7-adcc-ea26fe748a8e. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 890.716718] env[62753]: DEBUG nova.network.neutron [req-983db690-6c81-4d86-bfd2-5547da2e7fb9 req-75b53db4-40ba-4447-912f-5c1e595dcc86 service nova] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Updating instance_info_cache with network_info: [{"id": "1f18d3a5-8537-41a7-adcc-ea26fe748a8e", "address": "fa:16:3e:54:35:11", "network": {"id": "01ef394e-a94d-4d7e-bfe7-4d73f44c8320", "bridge": "br-int", "label": "tempest-ServerMetadataNegativeTestJSON-383531650-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "f0fb90849a6c494996e4fba4ea404cc4", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c894ab55-c869-4530-9702-cb46d173ce94", "external-id": "nsx-vlan-transportzone-792", "segmentation_id": 792, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap1f18d3a5-85", "ovs_interfaceid": "1f18d3a5-8537-41a7-adcc-ea26fe748a8e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 890.717888] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "refresh_cache-f03739f4-921f-46e5-b0e7-e9a99d96d621" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 890.718058] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "refresh_cache-f03739f4-921f-46e5-b0e7-e9a99d96d621" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 890.718227] env[62753]: DEBUG nova.network.neutron [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 890.756560] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a1a8569-bf3a-4b25-ae9f-6e9e7573c971 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "33c1de85-aade-4c69-8fdb-7672d8c53300" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 60.696s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 890.852531] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52475f92-57c9-425f-d599-8ca631a9c593, 'name': SearchDatastore_Task, 'duration_secs': 0.02571} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 890.853426] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1fc4b810-ddad-46d0-b54a-e369350873b0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 890.859396] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Waiting for the task: (returnval){ [ 890.859396] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d03b34-ee9f-6557-101f-00cff9fb8a86" [ 890.859396] env[62753]: _type = "Task" [ 890.859396] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 890.867210] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d03b34-ee9f-6557-101f-00cff9fb8a86, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.006364] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "b45bfcb4-b2cd-468c-9595-62c155d520fb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.006608] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "b45bfcb4-b2cd-468c-9595-62c155d520fb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.021115] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquiring lock "49311ef5-733a-4146-87ac-876a121a8d4d" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.021425] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "49311ef5-733a-4146-87ac-876a121a8d4d" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.021676] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquiring lock "49311ef5-733a-4146-87ac-876a121a8d4d-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 891.021900] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "49311ef5-733a-4146-87ac-876a121a8d4d-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 891.022143] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "49311ef5-733a-4146-87ac-876a121a8d4d-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 891.024835] env[62753]: INFO nova.compute.manager [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Terminating instance [ 891.026781] env[62753]: DEBUG nova.compute.manager [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 891.027013] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 891.027927] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0981194-6656-448d-bf53-7e7c3561ed34 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.036132] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 891.036132] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-67ac312a-0419-4552-9235-df91b2708774 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.042334] env[62753]: DEBUG oslo_vmware.api [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 891.042334] env[62753]: value = "task-1332286" [ 891.042334] env[62753]: _type = "Task" [ 891.042334] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.050692] env[62753]: DEBUG oslo_vmware.api [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332286, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.220925] env[62753]: DEBUG oslo_concurrency.lockutils [req-983db690-6c81-4d86-bfd2-5547da2e7fb9 req-75b53db4-40ba-4447-912f-5c1e595dcc86 service nova] Releasing lock "refresh_cache-57eef548-a0f1-4f0b-a026-885a10d005c2" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.221284] env[62753]: DEBUG nova.compute.manager [req-983db690-6c81-4d86-bfd2-5547da2e7fb9 req-75b53db4-40ba-4447-912f-5c1e595dcc86 service nova] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Received event network-changed-e44dcb1f-60ee-409a-805b-0c795852f967 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 891.221484] env[62753]: DEBUG nova.compute.manager [req-983db690-6c81-4d86-bfd2-5547da2e7fb9 req-75b53db4-40ba-4447-912f-5c1e595dcc86 service nova] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Refreshing instance network info cache due to event network-changed-e44dcb1f-60ee-409a-805b-0c795852f967. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 891.221705] env[62753]: DEBUG oslo_concurrency.lockutils [req-983db690-6c81-4d86-bfd2-5547da2e7fb9 req-75b53db4-40ba-4447-912f-5c1e595dcc86 service nova] Acquiring lock "refresh_cache-c5dae9a3-052b-4f4c-ac79-84aeaded457a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 891.221852] env[62753]: DEBUG oslo_concurrency.lockutils [req-983db690-6c81-4d86-bfd2-5547da2e7fb9 req-75b53db4-40ba-4447-912f-5c1e595dcc86 service nova] Acquired lock "refresh_cache-c5dae9a3-052b-4f4c-ac79-84aeaded457a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 891.222025] env[62753]: DEBUG nova.network.neutron [req-983db690-6c81-4d86-bfd2-5547da2e7fb9 req-75b53db4-40ba-4447-912f-5c1e595dcc86 service nova] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Refreshing network info cache for port e44dcb1f-60ee-409a-805b-0c795852f967 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 891.254543] env[62753]: DEBUG nova.network.neutron [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 891.370409] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d03b34-ee9f-6557-101f-00cff9fb8a86, 'name': SearchDatastore_Task, 'duration_secs': 0.010728} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.370672] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.370968] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 57eef548-a0f1-4f0b-a026-885a10d005c2/57eef548-a0f1-4f0b-a026-885a10d005c2.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 891.371268] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b4fda494-fce6-4efd-85a7-0dfd88d77389 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.379401] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Waiting for the task: (returnval){ [ 891.379401] env[62753]: value = "task-1332287" [ 891.379401] env[62753]: _type = "Task" [ 891.379401] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.387525] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332287, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.456834] env[62753]: DEBUG nova.network.neutron [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Updating instance_info_cache with network_info: [{"id": "ce484d8b-65ec-4512-aaf1-48a0c51f4f74", "address": "fa:16:3e:48:93:d8", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce484d8b-65", "ovs_interfaceid": "ce484d8b-65ec-4512-aaf1-48a0c51f4f74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 891.509376] env[62753]: DEBUG nova.compute.manager [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 891.551689] env[62753]: DEBUG oslo_vmware.api [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332286, 'name': PowerOffVM_Task, 'duration_secs': 0.486304} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 891.551974] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 891.552168] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 891.552426] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8a40819e-ec4c-43e8-b731-9ee00999949c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.630941] env[62753]: DEBUG nova.compute.manager [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 891.652147] env[62753]: DEBUG nova.virt.hardware [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 891.652416] env[62753]: DEBUG nova.virt.hardware [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 891.652576] env[62753]: DEBUG nova.virt.hardware [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 891.652759] env[62753]: DEBUG nova.virt.hardware [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 891.652911] env[62753]: DEBUG nova.virt.hardware [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 891.653077] env[62753]: DEBUG nova.virt.hardware [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 891.653295] env[62753]: DEBUG nova.virt.hardware [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 891.653456] env[62753]: DEBUG nova.virt.hardware [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 891.653624] env[62753]: DEBUG nova.virt.hardware [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 891.653785] env[62753]: DEBUG nova.virt.hardware [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 891.653959] env[62753]: DEBUG nova.virt.hardware [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 891.654861] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f72f91e-b46d-4424-9ee1-4f6a4ac970a1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.665114] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-330d558c-19de-430c-b0b8-7fd4c6e23001 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.693581] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 891.693787] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 891.694042] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Deleting the datastore file [datastore1] 49311ef5-733a-4146-87ac-876a121a8d4d {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 891.694321] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-720bd6d6-b011-4d32-9b47-b8accf7defb6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.701696] env[62753]: DEBUG oslo_vmware.api [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 891.701696] env[62753]: value = "task-1332289" [ 891.701696] env[62753]: _type = "Task" [ 891.701696] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 891.714511] env[62753]: DEBUG oslo_vmware.api [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332289, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.774174] env[62753]: INFO nova.compute.manager [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Rebuilding instance [ 891.834792] env[62753]: DEBUG nova.compute.manager [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 891.835736] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d35e9cc-3683-48cf-8ea8-60eb183be5a7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.892475] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332287, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 891.963027] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "refresh_cache-f03739f4-921f-46e5-b0e7-e9a99d96d621" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 891.963027] env[62753]: DEBUG nova.compute.manager [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Instance network_info: |[{"id": "ce484d8b-65ec-4512-aaf1-48a0c51f4f74", "address": "fa:16:3e:48:93:d8", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce484d8b-65", "ovs_interfaceid": "ce484d8b-65ec-4512-aaf1-48a0c51f4f74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 891.963027] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:48:93:d8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a10c88d7-d13f-44fd-acee-7a734eb5f56a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ce484d8b-65ec-4512-aaf1-48a0c51f4f74', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 891.970128] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Creating folder: Project (1d52d6eaee934be5ab0e0003df1ce316). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 891.974138] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9d6b9039-27e3-464e-9126-c0e15c307aa9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.989249] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Created folder: Project (1d52d6eaee934be5ab0e0003df1ce316) in parent group-v284541. [ 891.989686] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Creating folder: Instances. Parent ref: group-v284598. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 891.991087] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2253f33a-059c-4226-ad20-e333556aabf1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 891.994271] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f45adc10-5f4b-4453-9cf4-d97c0afab2fb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.006091] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a828c552-34d8-41ec-a9fd-558c8cba4e55 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.012019] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Created folder: Instances in parent group-v284598. [ 892.012019] env[62753]: DEBUG oslo.service.loopingcall [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.012019] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 892.012019] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c4827c43-aac2-4851-8d19-8bafd9b4cece {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.032333] env[62753]: DEBUG nova.network.neutron [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Successfully updated port: 61fe245f-aaa8-4566-857d-3002a5c3426d {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 892.059497] env[62753]: DEBUG nova.network.neutron [req-983db690-6c81-4d86-bfd2-5547da2e7fb9 req-75b53db4-40ba-4447-912f-5c1e595dcc86 service nova] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Updated VIF entry in instance network info cache for port e44dcb1f-60ee-409a-805b-0c795852f967. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 892.060174] env[62753]: DEBUG nova.network.neutron [req-983db690-6c81-4d86-bfd2-5547da2e7fb9 req-75b53db4-40ba-4447-912f-5c1e595dcc86 service nova] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Updating instance_info_cache with network_info: [{"id": "e44dcb1f-60ee-409a-805b-0c795852f967", "address": "fa:16:3e:0d:48:c3", "network": {"id": "c07a7726-4e7d-4f78-bdc8-ca9039c7997f", "bridge": "br-int", "label": "tempest-ServersTestBootFromVolume-687128365-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.171", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "a83c3627d2e040709b4d2d747758776b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8bb798a9-4c7f-4361-9436-a5690283861a", "external-id": "nsx-vlan-transportzone-547", "segmentation_id": 547, "bound_drivers": {"0": "nsxv3"}}, "devname": "tape44dcb1f-60", "ovs_interfaceid": "e44dcb1f-60ee-409a-805b-0c795852f967", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.064721] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Acquiring lock "refresh_cache-9a8a148c-057d-4c91-8b7c-4e05e6ac0f66" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.064989] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Acquired lock "refresh_cache-9a8a148c-057d-4c91-8b7c-4e05e6ac0f66" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.065269] env[62753]: DEBUG nova.network.neutron [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 892.069369] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bceba85-b28b-4992-800e-7694ade097fa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.073559] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 892.073559] env[62753]: value = "task-1332292" [ 892.073559] env[62753]: _type = "Task" [ 892.073559] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.081345] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e21cfe7b-05b7-4081-ac3a-09184c0f382d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.089321] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.089863] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332292, 'name': CreateVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.104783] env[62753]: DEBUG nova.compute.provider_tree [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 892.109865] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "interface-3bc05fdb-81be-4764-b9d9-01acd125b020-None" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.110108] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "interface-3bc05fdb-81be-4764-b9d9-01acd125b020-None" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.110419] env[62753]: DEBUG nova.objects.instance [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lazy-loading 'flavor' on Instance uuid 3bc05fdb-81be-4764-b9d9-01acd125b020 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.213880] env[62753]: DEBUG oslo_vmware.api [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332289, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.235362] env[62753]: DEBUG nova.compute.manager [req-9e5485ac-bc92-4f86-ba2f-32f03c5e45b2 req-bb36e074-6c26-480a-974d-1259fa1977bc service nova] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Received event network-vif-plugged-61fe245f-aaa8-4566-857d-3002a5c3426d {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.235643] env[62753]: DEBUG oslo_concurrency.lockutils [req-9e5485ac-bc92-4f86-ba2f-32f03c5e45b2 req-bb36e074-6c26-480a-974d-1259fa1977bc service nova] Acquiring lock "9a8a148c-057d-4c91-8b7c-4e05e6ac0f66-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.235853] env[62753]: DEBUG oslo_concurrency.lockutils [req-9e5485ac-bc92-4f86-ba2f-32f03c5e45b2 req-bb36e074-6c26-480a-974d-1259fa1977bc service nova] Lock "9a8a148c-057d-4c91-8b7c-4e05e6ac0f66-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.236118] env[62753]: DEBUG oslo_concurrency.lockutils [req-9e5485ac-bc92-4f86-ba2f-32f03c5e45b2 req-bb36e074-6c26-480a-974d-1259fa1977bc service nova] Lock "9a8a148c-057d-4c91-8b7c-4e05e6ac0f66-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.236275] env[62753]: DEBUG nova.compute.manager [req-9e5485ac-bc92-4f86-ba2f-32f03c5e45b2 req-bb36e074-6c26-480a-974d-1259fa1977bc service nova] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] No waiting events found dispatching network-vif-plugged-61fe245f-aaa8-4566-857d-3002a5c3426d {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 892.236448] env[62753]: WARNING nova.compute.manager [req-9e5485ac-bc92-4f86-ba2f-32f03c5e45b2 req-bb36e074-6c26-480a-974d-1259fa1977bc service nova] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Received unexpected event network-vif-plugged-61fe245f-aaa8-4566-857d-3002a5c3426d for instance with vm_state building and task_state spawning. [ 892.252529] env[62753]: DEBUG nova.compute.manager [req-02fd16dd-486a-4cc0-a166-ebff5d88473d req-2d0f433e-2335-4f11-9e42-0447d953f5e9 service nova] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Received event network-vif-plugged-ce484d8b-65ec-4512-aaf1-48a0c51f4f74 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.252862] env[62753]: DEBUG oslo_concurrency.lockutils [req-02fd16dd-486a-4cc0-a166-ebff5d88473d req-2d0f433e-2335-4f11-9e42-0447d953f5e9 service nova] Acquiring lock "f03739f4-921f-46e5-b0e7-e9a99d96d621-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 892.253064] env[62753]: DEBUG oslo_concurrency.lockutils [req-02fd16dd-486a-4cc0-a166-ebff5d88473d req-2d0f433e-2335-4f11-9e42-0447d953f5e9 service nova] Lock "f03739f4-921f-46e5-b0e7-e9a99d96d621-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 892.253263] env[62753]: DEBUG oslo_concurrency.lockutils [req-02fd16dd-486a-4cc0-a166-ebff5d88473d req-2d0f433e-2335-4f11-9e42-0447d953f5e9 service nova] Lock "f03739f4-921f-46e5-b0e7-e9a99d96d621-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 892.253492] env[62753]: DEBUG nova.compute.manager [req-02fd16dd-486a-4cc0-a166-ebff5d88473d req-2d0f433e-2335-4f11-9e42-0447d953f5e9 service nova] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] No waiting events found dispatching network-vif-plugged-ce484d8b-65ec-4512-aaf1-48a0c51f4f74 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 892.253591] env[62753]: WARNING nova.compute.manager [req-02fd16dd-486a-4cc0-a166-ebff5d88473d req-2d0f433e-2335-4f11-9e42-0447d953f5e9 service nova] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Received unexpected event network-vif-plugged-ce484d8b-65ec-4512-aaf1-48a0c51f4f74 for instance with vm_state building and task_state spawning. [ 892.253752] env[62753]: DEBUG nova.compute.manager [req-02fd16dd-486a-4cc0-a166-ebff5d88473d req-2d0f433e-2335-4f11-9e42-0447d953f5e9 service nova] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Received event network-changed-ce484d8b-65ec-4512-aaf1-48a0c51f4f74 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 892.254039] env[62753]: DEBUG nova.compute.manager [req-02fd16dd-486a-4cc0-a166-ebff5d88473d req-2d0f433e-2335-4f11-9e42-0447d953f5e9 service nova] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Refreshing instance network info cache due to event network-changed-ce484d8b-65ec-4512-aaf1-48a0c51f4f74. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 892.254172] env[62753]: DEBUG oslo_concurrency.lockutils [req-02fd16dd-486a-4cc0-a166-ebff5d88473d req-2d0f433e-2335-4f11-9e42-0447d953f5e9 service nova] Acquiring lock "refresh_cache-f03739f4-921f-46e5-b0e7-e9a99d96d621" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 892.254312] env[62753]: DEBUG oslo_concurrency.lockutils [req-02fd16dd-486a-4cc0-a166-ebff5d88473d req-2d0f433e-2335-4f11-9e42-0447d953f5e9 service nova] Acquired lock "refresh_cache-f03739f4-921f-46e5-b0e7-e9a99d96d621" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 892.254474] env[62753]: DEBUG nova.network.neutron [req-02fd16dd-486a-4cc0-a166-ebff5d88473d req-2d0f433e-2335-4f11-9e42-0447d953f5e9 service nova] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Refreshing network info cache for port ce484d8b-65ec-4512-aaf1-48a0c51f4f74 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 892.350525] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 892.350838] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ba8302ae-63ca-40bb-8a31-7a65ac1eeaed {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.357629] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 892.357629] env[62753]: value = "task-1332293" [ 892.357629] env[62753]: _type = "Task" [ 892.357629] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.366285] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332293, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.391313] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332287, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.71022} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.391569] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 57eef548-a0f1-4f0b-a026-885a10d005c2/57eef548-a0f1-4f0b-a026-885a10d005c2.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 892.391782] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 892.392044] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5139271f-5db7-4007-8ae9-f5d66ff5a8de {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.398945] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Waiting for the task: (returnval){ [ 892.398945] env[62753]: value = "task-1332294" [ 892.398945] env[62753]: _type = "Task" [ 892.398945] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.406934] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332294, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.568958] env[62753]: DEBUG oslo_concurrency.lockutils [req-983db690-6c81-4d86-bfd2-5547da2e7fb9 req-75b53db4-40ba-4447-912f-5c1e595dcc86 service nova] Releasing lock "refresh_cache-c5dae9a3-052b-4f4c-ac79-84aeaded457a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 892.585268] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332292, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.608368] env[62753]: DEBUG nova.scheduler.client.report [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 892.622957] env[62753]: DEBUG nova.network.neutron [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 892.683119] env[62753]: DEBUG nova.objects.instance [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lazy-loading 'pci_requests' on Instance uuid 3bc05fdb-81be-4764-b9d9-01acd125b020 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 892.714499] env[62753]: DEBUG oslo_vmware.api [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332289, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.764122} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.716584] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 892.716784] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 892.716960] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 892.717153] env[62753]: INFO nova.compute.manager [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Took 1.69 seconds to destroy the instance on the hypervisor. [ 892.717444] env[62753]: DEBUG oslo.service.loopingcall [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 892.717586] env[62753]: DEBUG nova.compute.manager [-] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 892.717678] env[62753]: DEBUG nova.network.neutron [-] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 892.822911] env[62753]: DEBUG nova.network.neutron [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Updating instance_info_cache with network_info: [{"id": "61fe245f-aaa8-4566-857d-3002a5c3426d", "address": "fa:16:3e:cf:4d:cd", "network": {"id": "b6b11d30-459c-4528-a5ae-de2c9c2bdeee", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-766685064-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "db8f96fcee6e47e786ceae2ed3c0645b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61fe245f-aa", "ovs_interfaceid": "61fe245f-aaa8-4566-857d-3002a5c3426d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 892.869737] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332293, 'name': PowerOffVM_Task, 'duration_secs': 0.126642} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.870070] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 892.870364] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 892.871428] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95576256-8340-4ed6-9923-f0c06b177e0a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.882361] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 892.882686] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f30cac3-04ae-46ad-81bf-d312a014aa21 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.908720] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 892.908938] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 892.909137] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Deleting the datastore file [datastore2] 33c1de85-aade-4c69-8fdb-7672d8c53300 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 892.912094] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-21b0af6f-7a49-4701-b3e0-1178c57ecd5e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.914173] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332294, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.096424} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 892.914704] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 892.915507] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aaa1e0fb-82ef-4ce9-94b6-625d6f17344d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.919033] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 892.919033] env[62753]: value = "task-1332296" [ 892.919033] env[62753]: _type = "Task" [ 892.919033] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.944592] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Reconfiguring VM instance instance-00000044 to attach disk [datastore1] 57eef548-a0f1-4f0b-a026-885a10d005c2/57eef548-a0f1-4f0b-a026-885a10d005c2.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 892.947876] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a1552dc-7674-4f99-a4c2-9156acaff1a4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 892.966306] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332296, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 892.971499] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Waiting for the task: (returnval){ [ 892.971499] env[62753]: value = "task-1332297" [ 892.971499] env[62753]: _type = "Task" [ 892.971499] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 892.979071] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332297, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.086494] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332292, 'name': CreateVM_Task, 'duration_secs': 0.822439} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.086494] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 893.087244] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.087374] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.087761] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 893.088055] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fe85f638-db4b-4493-bb96-414f9bfbb042 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.093419] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 893.093419] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e10249-95ee-6a3e-e2a4-50ea39a5a8cc" [ 893.093419] env[62753]: _type = "Task" [ 893.093419] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.101892] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e10249-95ee-6a3e-e2a4-50ea39a5a8cc, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.113279] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.498s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 893.113790] env[62753]: DEBUG nova.compute.manager [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 893.116575] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.200s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 893.116801] env[62753]: DEBUG nova.objects.instance [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Lazy-loading 'resources' on Instance uuid f132230d-0fba-4293-b940-85cc7ef1cad1 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 893.137830] env[62753]: DEBUG nova.network.neutron [req-02fd16dd-486a-4cc0-a166-ebff5d88473d req-2d0f433e-2335-4f11-9e42-0447d953f5e9 service nova] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Updated VIF entry in instance network info cache for port ce484d8b-65ec-4512-aaf1-48a0c51f4f74. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 893.138339] env[62753]: DEBUG nova.network.neutron [req-02fd16dd-486a-4cc0-a166-ebff5d88473d req-2d0f433e-2335-4f11-9e42-0447d953f5e9 service nova] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Updating instance_info_cache with network_info: [{"id": "ce484d8b-65ec-4512-aaf1-48a0c51f4f74", "address": "fa:16:3e:48:93:d8", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapce484d8b-65", "ovs_interfaceid": "ce484d8b-65ec-4512-aaf1-48a0c51f4f74", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 893.185815] env[62753]: DEBUG nova.objects.base [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Object Instance<3bc05fdb-81be-4764-b9d9-01acd125b020> lazy-loaded attributes: flavor,pci_requests {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 893.186093] env[62753]: DEBUG nova.network.neutron [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 893.248271] env[62753]: DEBUG nova.policy [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9931b40cfd7846038805c6a4caedaac0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b8e81660b30408c998e412f5fa81469', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 893.327640] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Releasing lock "refresh_cache-9a8a148c-057d-4c91-8b7c-4e05e6ac0f66" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.327989] env[62753]: DEBUG nova.compute.manager [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Instance network_info: |[{"id": "61fe245f-aaa8-4566-857d-3002a5c3426d", "address": "fa:16:3e:cf:4d:cd", "network": {"id": "b6b11d30-459c-4528-a5ae-de2c9c2bdeee", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-766685064-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "db8f96fcee6e47e786ceae2ed3c0645b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61fe245f-aa", "ovs_interfaceid": "61fe245f-aaa8-4566-857d-3002a5c3426d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 893.328458] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:cf:4d:cd', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cba18f15-a919-422e-a423-1e705e233389', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '61fe245f-aaa8-4566-857d-3002a5c3426d', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 893.336233] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Creating folder: Project (db8f96fcee6e47e786ceae2ed3c0645b). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 893.336524] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-d2f551cc-c526-4749-8ee7-9957969d27c9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.348481] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Created folder: Project (db8f96fcee6e47e786ceae2ed3c0645b) in parent group-v284541. [ 893.348678] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Creating folder: Instances. Parent ref: group-v284601. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 893.348912] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-30bd3b3f-f450-4de1-839d-b1f7aa657bad {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.361485] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Created folder: Instances in parent group-v284601. [ 893.361733] env[62753]: DEBUG oslo.service.loopingcall [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 893.361927] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 893.362158] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-ef9b9982-072c-42f8-a2af-67e6f415734c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.381778] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 893.381778] env[62753]: value = "task-1332300" [ 893.381778] env[62753]: _type = "Task" [ 893.381778] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.390538] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332300, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.428267] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332296, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.350368} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.428526] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 893.428715] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 893.428896] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 893.481352] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332297, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.604274] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e10249-95ee-6a3e-e2a4-50ea39a5a8cc, 'name': SearchDatastore_Task, 'duration_secs': 0.014706} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.604600] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.604836] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 893.605079] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 893.605236] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 893.605418] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 893.605684] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ad979cd7-d3e2-4cbd-aa39-923d61286384 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.615031] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 893.615031] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 893.615031] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-de1be93d-8add-45c2-9d01-0aafe0eea764 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.619622] env[62753]: DEBUG nova.compute.utils [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 893.626900] env[62753]: DEBUG nova.compute.manager [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 893.627122] env[62753]: DEBUG nova.network.neutron [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 893.628953] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 893.628953] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525cc938-0ed6-3284-bb46-ec710235bef3" [ 893.628953] env[62753]: _type = "Task" [ 893.628953] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.637955] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525cc938-0ed6-3284-bb46-ec710235bef3, 'name': SearchDatastore_Task, 'duration_secs': 0.010459} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.639355] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bb7f6cc-1b92-4935-875a-18c8fbc019f3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.645370] env[62753]: DEBUG oslo_concurrency.lockutils [req-02fd16dd-486a-4cc0-a166-ebff5d88473d req-2d0f433e-2335-4f11-9e42-0447d953f5e9 service nova] Releasing lock "refresh_cache-f03739f4-921f-46e5-b0e7-e9a99d96d621" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.656875] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 893.656875] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52fb06eb-5d2d-979a-8e0e-183cd6691cc2" [ 893.656875] env[62753]: _type = "Task" [ 893.656875] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.669178] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52fb06eb-5d2d-979a-8e0e-183cd6691cc2, 'name': SearchDatastore_Task, 'duration_secs': 0.009688} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.670557] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 893.670722] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] f03739f4-921f-46e5-b0e7-e9a99d96d621/f03739f4-921f-46e5-b0e7-e9a99d96d621.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 893.671372] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-116d2591-c337-4bb0-8e2c-da32c51ae746 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.677504] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 893.677504] env[62753]: value = "task-1332301" [ 893.677504] env[62753]: _type = "Task" [ 893.677504] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 893.678832] env[62753]: DEBUG nova.policy [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4dbc63095b0464fa0d2de86e3cf170e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5636da907ab343f9b42dbbd903d32283', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 893.686722] env[62753]: DEBUG nova.network.neutron [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Successfully created port: 9713beba-767d-408c-bd7d-bef58dc3c405 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 893.696809] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332301, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.897447] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332300, 'name': CreateVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 893.991389] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332297, 'name': ReconfigVM_Task, 'duration_secs': 0.945239} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 893.992519] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Reconfigured VM instance instance-00000044 to attach disk [datastore1] 57eef548-a0f1-4f0b-a026-885a10d005c2/57eef548-a0f1-4f0b-a026-885a10d005c2.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 893.993741] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4bae061-038b-4460-9c1b-181d20ab2d75 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 893.996961] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-11c79e56-0436-4917-bfac-d82c82bba845 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.004332] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba478dd-c371-4a39-b245-8cca45290026 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.009210] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Waiting for the task: (returnval){ [ 894.009210] env[62753]: value = "task-1332302" [ 894.009210] env[62753]: _type = "Task" [ 894.009210] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.045903] env[62753]: DEBUG nova.network.neutron [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Successfully created port: 66ec9e7b-4a31-4ee0-9ae0-f1ecf62c57bf {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 894.049040] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-643555f3-5017-467b-b0a4-7cd7ba4d5c7d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.057915] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332302, 'name': Rename_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.062234] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86d52b26-569b-45f8-8b02-aa5364dadb50 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.076065] env[62753]: DEBUG nova.compute.provider_tree [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 894.127872] env[62753]: DEBUG nova.compute.manager [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 894.191025] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332301, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.253822] env[62753]: DEBUG nova.network.neutron [-] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 894.282454] env[62753]: DEBUG nova.compute.manager [req-14c5ace5-247e-4ac2-b855-44839e70967a req-53bea3be-5fb5-4f0f-8b9c-35f276b2af37 service nova] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Received event network-changed-61fe245f-aaa8-4566-857d-3002a5c3426d {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 894.282454] env[62753]: DEBUG nova.compute.manager [req-14c5ace5-247e-4ac2-b855-44839e70967a req-53bea3be-5fb5-4f0f-8b9c-35f276b2af37 service nova] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Refreshing instance network info cache due to event network-changed-61fe245f-aaa8-4566-857d-3002a5c3426d. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 894.282778] env[62753]: DEBUG oslo_concurrency.lockutils [req-14c5ace5-247e-4ac2-b855-44839e70967a req-53bea3be-5fb5-4f0f-8b9c-35f276b2af37 service nova] Acquiring lock "refresh_cache-9a8a148c-057d-4c91-8b7c-4e05e6ac0f66" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 894.283112] env[62753]: DEBUG oslo_concurrency.lockutils [req-14c5ace5-247e-4ac2-b855-44839e70967a req-53bea3be-5fb5-4f0f-8b9c-35f276b2af37 service nova] Acquired lock "refresh_cache-9a8a148c-057d-4c91-8b7c-4e05e6ac0f66" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 894.283325] env[62753]: DEBUG nova.network.neutron [req-14c5ace5-247e-4ac2-b855-44839e70967a req-53bea3be-5fb5-4f0f-8b9c-35f276b2af37 service nova] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Refreshing network info cache for port 61fe245f-aaa8-4566-857d-3002a5c3426d {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 894.394465] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332300, 'name': CreateVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.468912] env[62753]: DEBUG nova.virt.hardware [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 894.469204] env[62753]: DEBUG nova.virt.hardware [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 894.469366] env[62753]: DEBUG nova.virt.hardware [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 894.469579] env[62753]: DEBUG nova.virt.hardware [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 894.469694] env[62753]: DEBUG nova.virt.hardware [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 894.469847] env[62753]: DEBUG nova.virt.hardware [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 894.470150] env[62753]: DEBUG nova.virt.hardware [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 894.470382] env[62753]: DEBUG nova.virt.hardware [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 894.470574] env[62753]: DEBUG nova.virt.hardware [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 894.470836] env[62753]: DEBUG nova.virt.hardware [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 894.471107] env[62753]: DEBUG nova.virt.hardware [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 894.471988] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9c30f59-cb99-4507-b295-e5a2bea2f2b5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.480336] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ddd355e-0c17-468b-8154-f7acbff182bc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.494019] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Instance VIF info [] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 894.499767] env[62753]: DEBUG oslo.service.loopingcall [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 894.499992] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 894.500265] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d4d4a7c6-146c-47e7-8cfe-3084f11c6991 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.520360] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332302, 'name': Rename_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.521596] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 894.521596] env[62753]: value = "task-1332303" [ 894.521596] env[62753]: _type = "Task" [ 894.521596] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.528749] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332303, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.579034] env[62753]: DEBUG nova.scheduler.client.report [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 894.693109] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332301, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.594233} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 894.693386] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] f03739f4-921f-46e5-b0e7-e9a99d96d621/f03739f4-921f-46e5-b0e7-e9a99d96d621.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 894.693596] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 894.693867] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2b8b2d18-a76f-49b3-99e0-56ccee387c83 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 894.702017] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 894.702017] env[62753]: value = "task-1332304" [ 894.702017] env[62753]: _type = "Task" [ 894.702017] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 894.710815] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332304, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 894.756380] env[62753]: INFO nova.compute.manager [-] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Took 2.04 seconds to deallocate network for instance. [ 894.894791] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332300, 'name': CreateVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.004435] env[62753]: DEBUG nova.network.neutron [req-14c5ace5-247e-4ac2-b855-44839e70967a req-53bea3be-5fb5-4f0f-8b9c-35f276b2af37 service nova] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Updated VIF entry in instance network info cache for port 61fe245f-aaa8-4566-857d-3002a5c3426d. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 895.004737] env[62753]: DEBUG nova.network.neutron [req-14c5ace5-247e-4ac2-b855-44839e70967a req-53bea3be-5fb5-4f0f-8b9c-35f276b2af37 service nova] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Updating instance_info_cache with network_info: [{"id": "61fe245f-aaa8-4566-857d-3002a5c3426d", "address": "fa:16:3e:cf:4d:cd", "network": {"id": "b6b11d30-459c-4528-a5ae-de2c9c2bdeee", "bridge": "br-int", "label": "tempest-ServerAddressesNegativeTestJSON-766685064-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "db8f96fcee6e47e786ceae2ed3c0645b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cba18f15-a919-422e-a423-1e705e233389", "external-id": "nsx-vlan-transportzone-79", "segmentation_id": 79, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap61fe245f-aa", "ovs_interfaceid": "61fe245f-aaa8-4566-857d-3002a5c3426d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 895.022638] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332302, 'name': Rename_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.031077] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332303, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.085226] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.968s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.090372] env[62753]: DEBUG oslo_concurrency.lockutils [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 19.695s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.090968] env[62753]: INFO nova.compute.claims [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 895.114365] env[62753]: INFO nova.scheduler.client.report [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Deleted allocations for instance f132230d-0fba-4293-b940-85cc7ef1cad1 [ 895.138469] env[62753]: DEBUG nova.compute.manager [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 895.171824] env[62753]: DEBUG nova.virt.hardware [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 895.171824] env[62753]: DEBUG nova.virt.hardware [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 895.171824] env[62753]: DEBUG nova.virt.hardware [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 895.172057] env[62753]: DEBUG nova.virt.hardware [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 895.172162] env[62753]: DEBUG nova.virt.hardware [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 895.172355] env[62753]: DEBUG nova.virt.hardware [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 895.172603] env[62753]: DEBUG nova.virt.hardware [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 895.172778] env[62753]: DEBUG nova.virt.hardware [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 895.173019] env[62753]: DEBUG nova.virt.hardware [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 895.173273] env[62753]: DEBUG nova.virt.hardware [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 895.173481] env[62753]: DEBUG nova.virt.hardware [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 895.174399] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9dc5876-c327-46f8-b451-4737a68e3617 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.184112] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-808bef74-6bbd-47fa-87b4-afb1fab52ff0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.211509] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332304, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070102} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.211801] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 895.212679] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d67f3a-ee11-4d63-bc6b-31a908428726 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.234934] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] f03739f4-921f-46e5-b0e7-e9a99d96d621/f03739f4-921f-46e5-b0e7-e9a99d96d621.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 895.235345] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6f4f122c-0cdf-4eb2-8073-2c52c0f0193c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.257119] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 895.257119] env[62753]: value = "task-1332305" [ 895.257119] env[62753]: _type = "Task" [ 895.257119] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.265971] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.266274] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332305, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.299714] env[62753]: DEBUG nova.network.neutron [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Successfully updated port: 9713beba-767d-408c-bd7d-bef58dc3c405 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 895.394655] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332300, 'name': CreateVM_Task, 'duration_secs': 1.610004} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.394878] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 895.395521] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.395701] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.396026] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 895.396276] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2ee5c04-8166-423b-a9c9-a09a24e77462 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.400614] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Waiting for the task: (returnval){ [ 895.400614] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525aaab4-d53e-31b1-e3be-83aec6eb9995" [ 895.400614] env[62753]: _type = "Task" [ 895.400614] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.407902] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525aaab4-d53e-31b1-e3be-83aec6eb9995, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.507551] env[62753]: DEBUG oslo_concurrency.lockutils [req-14c5ace5-247e-4ac2-b855-44839e70967a req-53bea3be-5fb5-4f0f-8b9c-35f276b2af37 service nova] Releasing lock "refresh_cache-9a8a148c-057d-4c91-8b7c-4e05e6ac0f66" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.507897] env[62753]: DEBUG nova.compute.manager [req-14c5ace5-247e-4ac2-b855-44839e70967a req-53bea3be-5fb5-4f0f-8b9c-35f276b2af37 service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Received event network-vif-deleted-f07bc1fd-c64f-4525-b198-0d7cc81d6927 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 895.508163] env[62753]: DEBUG nova.compute.manager [req-14c5ace5-247e-4ac2-b855-44839e70967a req-53bea3be-5fb5-4f0f-8b9c-35f276b2af37 service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Received event network-vif-deleted-2b4dd6f0-5b5b-47ee-b7cb-ecafe1190758 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 895.508354] env[62753]: DEBUG nova.compute.manager [req-14c5ace5-247e-4ac2-b855-44839e70967a req-53bea3be-5fb5-4f0f-8b9c-35f276b2af37 service nova] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Received event network-vif-deleted-51f3d79a-315b-418b-ac51-a3b5f5c328f3 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 895.522603] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332302, 'name': Rename_Task, 'duration_secs': 1.196679} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.522724] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 895.525839] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7da8195b-791a-476a-b72d-2fb8a736bcef {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.532658] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332303, 'name': CreateVM_Task, 'duration_secs': 0.698387} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.534439] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 895.534858] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Waiting for the task: (returnval){ [ 895.534858] env[62753]: value = "task-1332306" [ 895.534858] env[62753]: _type = "Task" [ 895.534858] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.535351] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.535573] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.535939] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 895.536279] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a326bfc6-4df8-4e5f-8d87-6c25cc44c094 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.543493] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 895.543493] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d22410-4afa-34eb-a76d-cf9f32759720" [ 895.543493] env[62753]: _type = "Task" [ 895.543493] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.548225] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332306, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.558462] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d22410-4afa-34eb-a76d-cf9f32759720, 'name': SearchDatastore_Task, 'duration_secs': 0.011346} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.558843] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.559174] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.559541] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.559812] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.560038] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.560362] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b54fe101-e36d-464b-952e-6ba17d9a9909 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.568674] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.568915] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 895.569684] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc6d1a77-5cab-4304-9634-87b44eb5c25d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.575130] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 895.575130] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]520a0885-0f50-e5ea-7e41-9fa8d58ba02b" [ 895.575130] env[62753]: _type = "Task" [ 895.575130] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.583135] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]520a0885-0f50-e5ea-7e41-9fa8d58ba02b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.622576] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a4cbd83e-6fc4-4038-a159-58c39818a2fe tempest-ServerTagsTestJSON-1307253807 tempest-ServerTagsTestJSON-1307253807-project-member] Lock "f132230d-0fba-4293-b940-85cc7ef1cad1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 25.616s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.680232] env[62753]: DEBUG nova.compute.manager [req-e8e5dd28-51c5-4a9b-af21-205dd74542a9 req-d169a0ee-0e3d-4a35-9fee-d60ccd015d21 service nova] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Received event network-vif-plugged-66ec9e7b-4a31-4ee0-9ae0-f1ecf62c57bf {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 895.680459] env[62753]: DEBUG oslo_concurrency.lockutils [req-e8e5dd28-51c5-4a9b-af21-205dd74542a9 req-d169a0ee-0e3d-4a35-9fee-d60ccd015d21 service nova] Acquiring lock "6b9a08b0-3cb3-474a-8193-953e74252a53-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 895.680670] env[62753]: DEBUG oslo_concurrency.lockutils [req-e8e5dd28-51c5-4a9b-af21-205dd74542a9 req-d169a0ee-0e3d-4a35-9fee-d60ccd015d21 service nova] Lock "6b9a08b0-3cb3-474a-8193-953e74252a53-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 895.680880] env[62753]: DEBUG oslo_concurrency.lockutils [req-e8e5dd28-51c5-4a9b-af21-205dd74542a9 req-d169a0ee-0e3d-4a35-9fee-d60ccd015d21 service nova] Lock "6b9a08b0-3cb3-474a-8193-953e74252a53-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 895.681385] env[62753]: DEBUG nova.compute.manager [req-e8e5dd28-51c5-4a9b-af21-205dd74542a9 req-d169a0ee-0e3d-4a35-9fee-d60ccd015d21 service nova] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] No waiting events found dispatching network-vif-plugged-66ec9e7b-4a31-4ee0-9ae0-f1ecf62c57bf {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 895.681649] env[62753]: WARNING nova.compute.manager [req-e8e5dd28-51c5-4a9b-af21-205dd74542a9 req-d169a0ee-0e3d-4a35-9fee-d60ccd015d21 service nova] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Received unexpected event network-vif-plugged-66ec9e7b-4a31-4ee0-9ae0-f1ecf62c57bf for instance with vm_state building and task_state spawning. [ 895.736492] env[62753]: DEBUG nova.network.neutron [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Successfully updated port: 66ec9e7b-4a31-4ee0-9ae0-f1ecf62c57bf {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 895.766606] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332305, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 895.803567] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.803805] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.804052] env[62753]: DEBUG nova.network.neutron [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 895.911942] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525aaab4-d53e-31b1-e3be-83aec6eb9995, 'name': SearchDatastore_Task, 'duration_secs': 0.011617} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 895.912257] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 895.912668] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 895.912728] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 895.912920] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 895.913122] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 895.913405] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-fdb85c7d-be1b-4270-93a5-4c35a72b3d24 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.921537] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 895.921752] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 895.922531] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ec49fd8-f2be-44ba-8504-a6605ab1706c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 895.928418] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Waiting for the task: (returnval){ [ 895.928418] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52225fc2-3a87-1537-7dc7-7f0e92e6455c" [ 895.928418] env[62753]: _type = "Task" [ 895.928418] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 895.936816] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52225fc2-3a87-1537-7dc7-7f0e92e6455c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.047131] env[62753]: DEBUG oslo_vmware.api [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332306, 'name': PowerOnVM_Task, 'duration_secs': 0.445115} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.047373] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 896.047615] env[62753]: INFO nova.compute.manager [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Took 9.24 seconds to spawn the instance on the hypervisor. [ 896.047809] env[62753]: DEBUG nova.compute.manager [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 896.048608] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-946a0df4-aa48-4191-a6a8-6549469fd868 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.085529] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]520a0885-0f50-e5ea-7e41-9fa8d58ba02b, 'name': SearchDatastore_Task, 'duration_secs': 0.014784} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.086379] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7bca128d-fc66-4853-aaa7-0ffdc8919e76 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.092282] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 896.092282] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]527432d7-7e51-a6b2-7943-7567e2f19140" [ 896.092282] env[62753]: _type = "Task" [ 896.092282] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.103130] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]527432d7-7e51-a6b2-7943-7567e2f19140, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.239653] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "refresh_cache-6b9a08b0-3cb3-474a-8193-953e74252a53" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.239798] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "refresh_cache-6b9a08b0-3cb3-474a-8193-953e74252a53" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 896.239947] env[62753]: DEBUG nova.network.neutron [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 896.271900] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332305, 'name': ReconfigVM_Task, 'duration_secs': 0.541309} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.272775] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Reconfigured VM instance instance-00000045 to attach disk [datastore1] f03739f4-921f-46e5-b0e7-e9a99d96d621/f03739f4-921f-46e5-b0e7-e9a99d96d621.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 896.273664] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6614858c-4d55-43fa-a123-0ecb2863dcb4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.282646] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 896.282646] env[62753]: value = "task-1332307" [ 896.282646] env[62753]: _type = "Task" [ 896.282646] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.290915] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332307, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.352173] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61e41f13-f86a-4159-b401-6ba1f0ad3ae5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.357792] env[62753]: WARNING nova.network.neutron [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] 8a9463a1-a861-4ade-8483-640392be0a89 already exists in list: networks containing: ['8a9463a1-a861-4ade-8483-640392be0a89']. ignoring it [ 896.363588] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c2993c5-22fd-47bd-b2e7-4ef2dd35d989 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.368274] env[62753]: DEBUG nova.compute.manager [req-f1b4a58a-c0cb-4052-aefe-26aa9b3cd0ca req-dccfe363-2758-46db-be00-5ff0aff1eb26 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Received event network-vif-plugged-9713beba-767d-408c-bd7d-bef58dc3c405 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.368509] env[62753]: DEBUG oslo_concurrency.lockutils [req-f1b4a58a-c0cb-4052-aefe-26aa9b3cd0ca req-dccfe363-2758-46db-be00-5ff0aff1eb26 service nova] Acquiring lock "3bc05fdb-81be-4764-b9d9-01acd125b020-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 896.368747] env[62753]: DEBUG oslo_concurrency.lockutils [req-f1b4a58a-c0cb-4052-aefe-26aa9b3cd0ca req-dccfe363-2758-46db-be00-5ff0aff1eb26 service nova] Lock "3bc05fdb-81be-4764-b9d9-01acd125b020-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 896.368921] env[62753]: DEBUG oslo_concurrency.lockutils [req-f1b4a58a-c0cb-4052-aefe-26aa9b3cd0ca req-dccfe363-2758-46db-be00-5ff0aff1eb26 service nova] Lock "3bc05fdb-81be-4764-b9d9-01acd125b020-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 896.369123] env[62753]: DEBUG nova.compute.manager [req-f1b4a58a-c0cb-4052-aefe-26aa9b3cd0ca req-dccfe363-2758-46db-be00-5ff0aff1eb26 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] No waiting events found dispatching network-vif-plugged-9713beba-767d-408c-bd7d-bef58dc3c405 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 896.369305] env[62753]: WARNING nova.compute.manager [req-f1b4a58a-c0cb-4052-aefe-26aa9b3cd0ca req-dccfe363-2758-46db-be00-5ff0aff1eb26 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Received unexpected event network-vif-plugged-9713beba-767d-408c-bd7d-bef58dc3c405 for instance with vm_state active and task_state None. [ 896.369499] env[62753]: DEBUG nova.compute.manager [req-f1b4a58a-c0cb-4052-aefe-26aa9b3cd0ca req-dccfe363-2758-46db-be00-5ff0aff1eb26 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Received event network-changed-9713beba-767d-408c-bd7d-bef58dc3c405 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 896.369650] env[62753]: DEBUG nova.compute.manager [req-f1b4a58a-c0cb-4052-aefe-26aa9b3cd0ca req-dccfe363-2758-46db-be00-5ff0aff1eb26 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Refreshing instance network info cache due to event network-changed-9713beba-767d-408c-bd7d-bef58dc3c405. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 896.369822] env[62753]: DEBUG oslo_concurrency.lockutils [req-f1b4a58a-c0cb-4052-aefe-26aa9b3cd0ca req-dccfe363-2758-46db-be00-5ff0aff1eb26 service nova] Acquiring lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 896.400190] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c13f8d2-79e7-488b-bb5f-cea2f361abb5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.408013] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a0eb82e-6902-494a-9fc9-e0104d82c85c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.424907] env[62753]: DEBUG nova.compute.provider_tree [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 896.437958] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52225fc2-3a87-1537-7dc7-7f0e92e6455c, 'name': SearchDatastore_Task, 'duration_secs': 0.020547} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.441455] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-22b881f6-0284-40b2-a691-22761a935f5c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.448103] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Waiting for the task: (returnval){ [ 896.448103] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52033e42-8fd6-5ba8-5fe2-691d668d0633" [ 896.448103] env[62753]: _type = "Task" [ 896.448103] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.455802] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52033e42-8fd6-5ba8-5fe2-691d668d0633, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.565646] env[62753]: INFO nova.compute.manager [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Took 36.54 seconds to build instance. [ 896.603118] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]527432d7-7e51-a6b2-7943-7567e2f19140, 'name': SearchDatastore_Task, 'duration_secs': 0.009462} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.603400] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.603722] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 33c1de85-aade-4c69-8fdb-7672d8c53300/33c1de85-aade-4c69-8fdb-7672d8c53300.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 896.603990] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d6e2e6ed-bed8-4280-b5c5-fb29bfb44acb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.612034] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 896.612034] env[62753]: value = "task-1332308" [ 896.612034] env[62753]: _type = "Task" [ 896.612034] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.620016] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332308, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.783057] env[62753]: DEBUG nova.network.neutron [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 896.800103] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332307, 'name': Rename_Task, 'duration_secs': 0.152097} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.800443] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 896.800722] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0ab24efe-04ce-40e9-b9b5-f1bc307ab3a8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.810632] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 896.810632] env[62753]: value = "task-1332309" [ 896.810632] env[62753]: _type = "Task" [ 896.810632] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.825251] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332309, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.927945] env[62753]: DEBUG nova.scheduler.client.report [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 896.959931] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52033e42-8fd6-5ba8-5fe2-691d668d0633, 'name': SearchDatastore_Task, 'duration_secs': 0.010647} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 896.960319] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 896.960522] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66/9a8a148c-057d-4c91-8b7c-4e05e6ac0f66.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 896.960811] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8ca2f852-e789-475d-aa2e-27a6f0b3442f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 896.967775] env[62753]: DEBUG nova.network.neutron [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Updating instance_info_cache with network_info: [{"id": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "address": "fa:16:3e:1d:84:1b", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8e502a-7a", "ovs_interfaceid": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9713beba-767d-408c-bd7d-bef58dc3c405", "address": "fa:16:3e:da:85:43", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9713beba-76", "ovs_interfaceid": "9713beba-767d-408c-bd7d-bef58dc3c405", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 896.970512] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Waiting for the task: (returnval){ [ 896.970512] env[62753]: value = "task-1332310" [ 896.970512] env[62753]: _type = "Task" [ 896.970512] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 896.982341] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': task-1332310, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 896.989592] env[62753]: DEBUG nova.network.neutron [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Updating instance_info_cache with network_info: [{"id": "66ec9e7b-4a31-4ee0-9ae0-f1ecf62c57bf", "address": "fa:16:3e:8b:25:88", "network": {"id": "35ed0ec0-821f-45b6-bc2d-015626bfc0a2", "bridge": "br-int", "label": "tempest-ServersTestJSON-444726172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5636da907ab343f9b42dbbd903d32283", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66ec9e7b-4a", "ovs_interfaceid": "66ec9e7b-4a31-4ee0-9ae0-f1ecf62c57bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.068106] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2c667333-0a8d-4809-b3fb-a12f5590b718 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Lock "57eef548-a0f1-4f0b-a026-885a10d005c2" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 58.581s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.123589] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332308, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.326036] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332309, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.434048] env[62753]: DEBUG oslo_concurrency.lockutils [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.345s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 897.434763] env[62753]: DEBUG nova.compute.manager [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 897.437973] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 21.781s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 897.438327] env[62753]: DEBUG nova.objects.instance [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Lazy-loading 'resources' on Instance uuid 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 897.472181] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.472944] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 897.473218] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.473694] env[62753]: DEBUG oslo_concurrency.lockutils [req-f1b4a58a-c0cb-4052-aefe-26aa9b3cd0ca req-dccfe363-2758-46db-be00-5ff0aff1eb26 service nova] Acquired lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 897.473932] env[62753]: DEBUG nova.network.neutron [req-f1b4a58a-c0cb-4052-aefe-26aa9b3cd0ca req-dccfe363-2758-46db-be00-5ff0aff1eb26 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Refreshing network info cache for port 9713beba-767d-408c-bd7d-bef58dc3c405 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 897.476741] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9c8858f-e67e-4609-a351-faa38a9c78bc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.506380] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "refresh_cache-6b9a08b0-3cb3-474a-8193-953e74252a53" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 897.506548] env[62753]: DEBUG nova.compute.manager [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Instance network_info: |[{"id": "66ec9e7b-4a31-4ee0-9ae0-f1ecf62c57bf", "address": "fa:16:3e:8b:25:88", "network": {"id": "35ed0ec0-821f-45b6-bc2d-015626bfc0a2", "bridge": "br-int", "label": "tempest-ServersTestJSON-444726172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5636da907ab343f9b42dbbd903d32283", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66ec9e7b-4a", "ovs_interfaceid": "66ec9e7b-4a31-4ee0-9ae0-f1ecf62c57bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 897.507665] env[62753]: DEBUG nova.virt.hardware [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 897.507898] env[62753]: DEBUG nova.virt.hardware [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 897.508381] env[62753]: DEBUG nova.virt.hardware [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 897.508612] env[62753]: DEBUG nova.virt.hardware [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 897.508768] env[62753]: DEBUG nova.virt.hardware [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 897.509059] env[62753]: DEBUG nova.virt.hardware [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 897.509308] env[62753]: DEBUG nova.virt.hardware [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 897.509542] env[62753]: DEBUG nova.virt.hardware [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 897.510082] env[62753]: DEBUG nova.virt.hardware [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 897.510481] env[62753]: DEBUG nova.virt.hardware [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 897.510712] env[62753]: DEBUG nova.virt.hardware [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 897.519068] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Reconfiguring VM to attach interface {{(pid=62753) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 897.523257] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8b:25:88', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2624812a-9f9c-461d-8b5f-79bea90c7ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '66ec9e7b-4a31-4ee0-9ae0-f1ecf62c57bf', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 897.531216] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Creating folder: Project (5636da907ab343f9b42dbbd903d32283). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 897.532020] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b9d7766-02bd-440d-90b9-4984a27c9ecb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.545289] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': task-1332310, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.548015] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-bdcdfd17-ddf4-4408-86df-7478e78a0c80 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.558111] env[62753]: DEBUG oslo_vmware.api [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 897.558111] env[62753]: value = "task-1332311" [ 897.558111] env[62753]: _type = "Task" [ 897.558111] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.575011] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Created folder: Project (5636da907ab343f9b42dbbd903d32283) in parent group-v284541. [ 897.575011] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Creating folder: Instances. Parent ref: group-v284605. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 897.575011] env[62753]: DEBUG oslo_vmware.api [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332311, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.575438] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-878e3bd9-24e1-4708-8237-e7217cd29566 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.586790] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Created folder: Instances in parent group-v284605. [ 897.587069] env[62753]: DEBUG oslo.service.loopingcall [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 897.587335] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 897.587574] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0c499a5f-77b8-4bc2-b4ce-1e8276423847 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.609764] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 897.609764] env[62753]: value = "task-1332314" [ 897.609764] env[62753]: _type = "Task" [ 897.609764] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.621534] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332314, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.629560] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332308, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.515441} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.629904] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 33c1de85-aade-4c69-8fdb-7672d8c53300/33c1de85-aade-4c69-8fdb-7672d8c53300.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 897.630144] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 897.630398] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-64a78bf7-7034-4483-bd16-19cf59aef6d6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.638862] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 897.638862] env[62753]: value = "task-1332315" [ 897.638862] env[62753]: _type = "Task" [ 897.638862] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 897.647754] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332315, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 897.771134] env[62753]: DEBUG nova.network.neutron [req-f1b4a58a-c0cb-4052-aefe-26aa9b3cd0ca req-dccfe363-2758-46db-be00-5ff0aff1eb26 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Updated VIF entry in instance network info cache for port 9713beba-767d-408c-bd7d-bef58dc3c405. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 897.771134] env[62753]: DEBUG nova.network.neutron [req-f1b4a58a-c0cb-4052-aefe-26aa9b3cd0ca req-dccfe363-2758-46db-be00-5ff0aff1eb26 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Updating instance_info_cache with network_info: [{"id": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "address": "fa:16:3e:1d:84:1b", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8e502a-7a", "ovs_interfaceid": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9713beba-767d-408c-bd7d-bef58dc3c405", "address": "fa:16:3e:da:85:43", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9713beba-76", "ovs_interfaceid": "9713beba-767d-408c-bd7d-bef58dc3c405", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 897.825590] env[62753]: DEBUG oslo_vmware.api [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332309, 'name': PowerOnVM_Task, 'duration_secs': 0.979576} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.826560] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 897.827105] env[62753]: INFO nova.compute.manager [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Took 8.60 seconds to spawn the instance on the hypervisor. [ 897.827464] env[62753]: DEBUG nova.compute.manager [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 897.829037] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c23c722-7327-4118-9dbd-2c63e56b943e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.944861] env[62753]: DEBUG nova.compute.utils [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 897.950143] env[62753]: DEBUG nova.compute.manager [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 897.950324] env[62753]: DEBUG nova.network.neutron [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 897.989454] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': task-1332310, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.669145} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 897.989721] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66/9a8a148c-057d-4c91-8b7c-4e05e6ac0f66.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 897.989941] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 897.991165] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1a0b239a-99c3-404d-8ec3-0c4a4d431612 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 897.998445] env[62753]: DEBUG nova.policy [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'af45e909b0c84662a8cfdf92896e5fe5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '196457eaa51643c38b4ef4fc6fbbcba3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 898.003530] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Waiting for the task: (returnval){ [ 898.003530] env[62753]: value = "task-1332316" [ 898.003530] env[62753]: _type = "Task" [ 898.003530] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.012936] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': task-1332316, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.035038] env[62753]: DEBUG nova.compute.manager [req-e1e58538-5c62-4165-a70a-4f5a865c1a5d req-5801961a-b2cd-49aa-b2d7-5908dfe25465 service nova] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Received event network-changed-66ec9e7b-4a31-4ee0-9ae0-f1ecf62c57bf {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 898.035728] env[62753]: DEBUG nova.compute.manager [req-e1e58538-5c62-4165-a70a-4f5a865c1a5d req-5801961a-b2cd-49aa-b2d7-5908dfe25465 service nova] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Refreshing instance network info cache due to event network-changed-66ec9e7b-4a31-4ee0-9ae0-f1ecf62c57bf. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 898.036091] env[62753]: DEBUG oslo_concurrency.lockutils [req-e1e58538-5c62-4165-a70a-4f5a865c1a5d req-5801961a-b2cd-49aa-b2d7-5908dfe25465 service nova] Acquiring lock "refresh_cache-6b9a08b0-3cb3-474a-8193-953e74252a53" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.036322] env[62753]: DEBUG oslo_concurrency.lockutils [req-e1e58538-5c62-4165-a70a-4f5a865c1a5d req-5801961a-b2cd-49aa-b2d7-5908dfe25465 service nova] Acquired lock "refresh_cache-6b9a08b0-3cb3-474a-8193-953e74252a53" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.036557] env[62753]: DEBUG nova.network.neutron [req-e1e58538-5c62-4165-a70a-4f5a865c1a5d req-5801961a-b2cd-49aa-b2d7-5908dfe25465 service nova] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Refreshing network info cache for port 66ec9e7b-4a31-4ee0-9ae0-f1ecf62c57bf {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 898.072021] env[62753]: DEBUG oslo_vmware.api [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332311, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.118746] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332314, 'name': CreateVM_Task, 'duration_secs': 0.36731} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.121255] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 898.122331] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.122542] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.122981] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 898.123342] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5b9ca62b-a36b-42e8-a7aa-87460971dc58 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.129074] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 898.129074] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a2f71f-378e-3c04-8bcd-dfa7842785cf" [ 898.129074] env[62753]: _type = "Task" [ 898.129074] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.139180] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a2f71f-378e-3c04-8bcd-dfa7842785cf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.146327] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332315, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.127068} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.148802] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 898.151153] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca0d5b55-1cf0-4f05-b313-7053b4851223 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.174284] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Reconfiguring VM instance instance-00000043 to attach disk [datastore2] 33c1de85-aade-4c69-8fdb-7672d8c53300/33c1de85-aade-4c69-8fdb-7672d8c53300.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 898.177574] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d4a2ef7a-512f-4407-926f-0e4664c65b41 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.199836] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 898.199836] env[62753]: value = "task-1332317" [ 898.199836] env[62753]: _type = "Task" [ 898.199836] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.210124] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332317, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.280969] env[62753]: DEBUG oslo_concurrency.lockutils [req-f1b4a58a-c0cb-4052-aefe-26aa9b3cd0ca req-dccfe363-2758-46db-be00-5ff0aff1eb26 service nova] Releasing lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.305613] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92f78d5d-f2c9-47d2-bc62-06f2f5ce5b53 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.314310] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4845308b-93f5-46a1-b575-339f0bd2fa76 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.358609] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e73b339-aae6-4bb2-a14f-301d78a7aaee {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.363089] env[62753]: INFO nova.compute.manager [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Took 35.52 seconds to build instance. [ 898.369017] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-170c975b-b149-4882-b28b-1ba806634e89 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.385558] env[62753]: DEBUG nova.compute.provider_tree [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 898.459850] env[62753]: DEBUG nova.compute.manager [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 898.487163] env[62753]: DEBUG nova.network.neutron [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Successfully created port: 81591716-fb4a-427f-b3e9-a8b265fe1ce1 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 898.515347] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': task-1332316, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070027} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.515625] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 898.516467] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-add5ecf7-e857-4c6b-9205-263a25f5d093 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.540071] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Reconfiguring VM instance instance-00000046 to attach disk [datastore1] 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66/9a8a148c-057d-4c91-8b7c-4e05e6ac0f66.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 898.543282] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-673a4701-6a31-4f50-a12f-9fa20d23dd9d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.564016] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Waiting for the task: (returnval){ [ 898.564016] env[62753]: value = "task-1332318" [ 898.564016] env[62753]: _type = "Task" [ 898.564016] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.571525] env[62753]: DEBUG oslo_vmware.api [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332311, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.576537] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': task-1332318, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.641527] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a2f71f-378e-3c04-8bcd-dfa7842785cf, 'name': SearchDatastore_Task, 'duration_secs': 0.042865} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 898.641896] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 898.642367] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 898.642674] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 898.642875] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 898.643198] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 898.643562] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3e911d24-df50-4b8d-b8bd-fa80137d2c8f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.652930] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 898.653222] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 898.654012] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a066ee36-be5e-45da-850a-055d89f01156 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 898.659408] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 898.659408] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523d6ed1-d859-4b53-d223-329876f44255" [ 898.659408] env[62753]: _type = "Task" [ 898.659408] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 898.667408] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523d6ed1-d859-4b53-d223-329876f44255, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.712869] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332317, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 898.868736] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9cfd12a3-1688-4478-a1fa-3eb754d43f73 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "f03739f4-921f-46e5-b0e7-e9a99d96d621" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 59.254s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 898.891832] env[62753]: DEBUG nova.scheduler.client.report [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 898.948393] env[62753]: DEBUG nova.network.neutron [req-e1e58538-5c62-4165-a70a-4f5a865c1a5d req-5801961a-b2cd-49aa-b2d7-5908dfe25465 service nova] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Updated VIF entry in instance network info cache for port 66ec9e7b-4a31-4ee0-9ae0-f1ecf62c57bf. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 898.948769] env[62753]: DEBUG nova.network.neutron [req-e1e58538-5c62-4165-a70a-4f5a865c1a5d req-5801961a-b2cd-49aa-b2d7-5908dfe25465 service nova] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Updating instance_info_cache with network_info: [{"id": "66ec9e7b-4a31-4ee0-9ae0-f1ecf62c57bf", "address": "fa:16:3e:8b:25:88", "network": {"id": "35ed0ec0-821f-45b6-bc2d-015626bfc0a2", "bridge": "br-int", "label": "tempest-ServersTestJSON-444726172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5636da907ab343f9b42dbbd903d32283", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap66ec9e7b-4a", "ovs_interfaceid": "66ec9e7b-4a31-4ee0-9ae0-f1ecf62c57bf", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 899.073462] env[62753]: DEBUG oslo_vmware.api [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332311, 'name': ReconfigVM_Task, 'duration_secs': 1.284056} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.076753] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.076998] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Reconfigured VM to attach interface {{(pid=62753) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 899.080647] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': task-1332318, 'name': ReconfigVM_Task, 'duration_secs': 0.415236} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.080910] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Reconfigured VM instance instance-00000046 to attach disk [datastore1] 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66/9a8a148c-057d-4c91-8b7c-4e05e6ac0f66.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 899.081632] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-26fa4bbb-5e90-4246-b285-48fc67132553 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.087408] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Waiting for the task: (returnval){ [ 899.087408] env[62753]: value = "task-1332319" [ 899.087408] env[62753]: _type = "Task" [ 899.087408] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.095903] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': task-1332319, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.169783] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523d6ed1-d859-4b53-d223-329876f44255, 'name': SearchDatastore_Task, 'duration_secs': 0.01281} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.171159] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8956aaf6-2d28-4e5c-a8b5-8644dc97bb67 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.176452] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 899.176452] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5284e3d5-8dc2-f4cc-cce2-16bdcaa2b214" [ 899.176452] env[62753]: _type = "Task" [ 899.176452] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.184623] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5284e3d5-8dc2-f4cc-cce2-16bdcaa2b214, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.209305] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332317, 'name': ReconfigVM_Task, 'duration_secs': 0.641747} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.209547] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Reconfigured VM instance instance-00000043 to attach disk [datastore2] 33c1de85-aade-4c69-8fdb-7672d8c53300/33c1de85-aade-4c69-8fdb-7672d8c53300.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 899.210163] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-61b381f6-4508-487d-bbc2-b899d41c463a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.216050] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 899.216050] env[62753]: value = "task-1332320" [ 899.216050] env[62753]: _type = "Task" [ 899.216050] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.223855] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332320, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.398348] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.960s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.400921] env[62753]: DEBUG oslo_concurrency.lockutils [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 22.299s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.401353] env[62753]: DEBUG nova.objects.instance [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lazy-loading 'resources' on Instance uuid 046fc3dc-a187-4fda-bc66-345e1226f83d {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 899.420937] env[62753]: INFO nova.scheduler.client.report [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Deleted allocations for instance 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3 [ 899.451825] env[62753]: DEBUG oslo_concurrency.lockutils [req-e1e58538-5c62-4165-a70a-4f5a865c1a5d req-5801961a-b2cd-49aa-b2d7-5908dfe25465 service nova] Releasing lock "refresh_cache-6b9a08b0-3cb3-474a-8193-953e74252a53" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.470920] env[62753]: DEBUG nova.compute.manager [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 899.496748] env[62753]: DEBUG nova.virt.hardware [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 899.497025] env[62753]: DEBUG nova.virt.hardware [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 899.497200] env[62753]: DEBUG nova.virt.hardware [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 899.497393] env[62753]: DEBUG nova.virt.hardware [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 899.497544] env[62753]: DEBUG nova.virt.hardware [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 899.497694] env[62753]: DEBUG nova.virt.hardware [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 899.497908] env[62753]: DEBUG nova.virt.hardware [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 899.498278] env[62753]: DEBUG nova.virt.hardware [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 899.498278] env[62753]: DEBUG nova.virt.hardware [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 899.498417] env[62753]: DEBUG nova.virt.hardware [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 899.498594] env[62753]: DEBUG nova.virt.hardware [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 899.499672] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ab849d-fc66-4a7e-8c8c-b3bbaa331c35 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.507879] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b2d5b051-372e-424c-8ee4-064e3f1eed80 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.537736] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "f03739f4-921f-46e5-b0e7-e9a99d96d621" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.537980] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "f03739f4-921f-46e5-b0e7-e9a99d96d621" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.581900] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7d0120d7-357d-4122-a521-b0093afd4122 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "interface-3bc05fdb-81be-4764-b9d9-01acd125b020-None" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 7.472s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.598070] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': task-1332319, 'name': Rename_Task, 'duration_secs': 0.3975} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.598331] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 899.598580] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ae038e59-c8fa-44a2-b174-25b4fd234149 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.605174] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Waiting for the task: (returnval){ [ 899.605174] env[62753]: value = "task-1332321" [ 899.605174] env[62753]: _type = "Task" [ 899.605174] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.614782] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': task-1332321, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.687303] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5284e3d5-8dc2-f4cc-cce2-16bdcaa2b214, 'name': SearchDatastore_Task, 'duration_secs': 0.041042} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 899.687509] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 899.687768] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 6b9a08b0-3cb3-474a-8193-953e74252a53/6b9a08b0-3cb3-474a-8193-953e74252a53.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 899.688046] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5ca00839-598a-4530-bd06-998ff97b12d0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.694108] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 899.694108] env[62753]: value = "task-1332322" [ 899.694108] env[62753]: _type = "Task" [ 899.694108] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 899.701790] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332322, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.725712] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332320, 'name': Rename_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 899.932589] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1561b495-51f7-432e-b8b5-f03329d104b9 tempest-ServerShowV254Test-1306584288 tempest-ServerShowV254Test-1306584288-project-member] Lock "4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 27.972s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.978942] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Acquiring lock "57eef548-a0f1-4f0b-a026-885a10d005c2" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.979225] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Lock "57eef548-a0f1-4f0b-a026-885a10d005c2" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.979436] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Acquiring lock "57eef548-a0f1-4f0b-a026-885a10d005c2-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 899.979620] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Lock "57eef548-a0f1-4f0b-a026-885a10d005c2-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 899.980761] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Lock "57eef548-a0f1-4f0b-a026-885a10d005c2-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 899.982191] env[62753]: INFO nova.compute.manager [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Terminating instance [ 899.983861] env[62753]: DEBUG nova.compute.manager [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 899.984081] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 899.984912] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5734d8d0-8f74-41a6-9bce-d900e8e29f7d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 899.997451] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 899.997451] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ecd6fa3f-7d0d-41d0-aa0d-15d54b1ce6d0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.007015] env[62753]: DEBUG oslo_vmware.api [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Waiting for the task: (returnval){ [ 900.007015] env[62753]: value = "task-1332323" [ 900.007015] env[62753]: _type = "Task" [ 900.007015] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.017691] env[62753]: DEBUG oslo_vmware.api [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332323, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.041568] env[62753]: DEBUG nova.compute.utils [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 900.069398] env[62753]: DEBUG nova.network.neutron [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Successfully updated port: 81591716-fb4a-427f-b3e9-a8b265fe1ce1 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 900.119057] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': task-1332321, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.206886] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332322, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.209366] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5490b654-d6e5-4cd0-9f32-4b4264a82065 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.222222] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dca2674f-a8ae-44ad-9946-33989fe69b9b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.232550] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332320, 'name': Rename_Task, 'duration_secs': 0.725519} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.260623] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 900.261228] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-be3f89af-8903-44fb-ae02-2ab47d014485 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.263626] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b8b2ec-2449-480d-be8a-acc263867f42 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.272031] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8577e8db-cae6-4662-a7d6-ad566e042a0e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.277549] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 900.277549] env[62753]: value = "task-1332324" [ 900.277549] env[62753]: _type = "Task" [ 900.277549] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.291809] env[62753]: DEBUG nova.compute.provider_tree [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 900.300810] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332324, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.517984] env[62753]: DEBUG oslo_vmware.api [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332323, 'name': PowerOffVM_Task, 'duration_secs': 0.229989} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.518365] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 900.518596] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 900.518915] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-bf5c2ffe-d3bf-415d-95be-01a0b460384f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.547758] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "f03739f4-921f-46e5-b0e7-e9a99d96d621" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.572344] env[62753]: DEBUG oslo_concurrency.lockutils [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 900.572534] env[62753]: DEBUG oslo_concurrency.lockutils [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 900.572752] env[62753]: DEBUG nova.network.neutron [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 900.575828] env[62753]: DEBUG nova.compute.manager [req-626d1292-3759-406b-bd5f-22699fb8a65f req-77766563-3a14-4600-9929-e24a6c040a97 service nova] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Received event network-vif-plugged-81591716-fb4a-427f-b3e9-a8b265fe1ce1 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 900.576065] env[62753]: DEBUG oslo_concurrency.lockutils [req-626d1292-3759-406b-bd5f-22699fb8a65f req-77766563-3a14-4600-9929-e24a6c040a97 service nova] Acquiring lock "c73b1ae0-dc98-47f7-babf-e96169384785-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 900.576279] env[62753]: DEBUG oslo_concurrency.lockutils [req-626d1292-3759-406b-bd5f-22699fb8a65f req-77766563-3a14-4600-9929-e24a6c040a97 service nova] Lock "c73b1ae0-dc98-47f7-babf-e96169384785-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 900.576456] env[62753]: DEBUG oslo_concurrency.lockutils [req-626d1292-3759-406b-bd5f-22699fb8a65f req-77766563-3a14-4600-9929-e24a6c040a97 service nova] Lock "c73b1ae0-dc98-47f7-babf-e96169384785-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 900.576672] env[62753]: DEBUG nova.compute.manager [req-626d1292-3759-406b-bd5f-22699fb8a65f req-77766563-3a14-4600-9929-e24a6c040a97 service nova] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] No waiting events found dispatching network-vif-plugged-81591716-fb4a-427f-b3e9-a8b265fe1ce1 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 900.576789] env[62753]: WARNING nova.compute.manager [req-626d1292-3759-406b-bd5f-22699fb8a65f req-77766563-3a14-4600-9929-e24a6c040a97 service nova] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Received unexpected event network-vif-plugged-81591716-fb4a-427f-b3e9-a8b265fe1ce1 for instance with vm_state building and task_state spawning. [ 900.600782] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 900.601163] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 900.601407] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Deleting the datastore file [datastore1] 57eef548-a0f1-4f0b-a026-885a10d005c2 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 900.601715] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c11df1ed-1094-4d30-90fc-e7e9d87f1e00 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.612366] env[62753]: DEBUG oslo_vmware.api [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Waiting for the task: (returnval){ [ 900.612366] env[62753]: value = "task-1332326" [ 900.612366] env[62753]: _type = "Task" [ 900.612366] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 900.618153] env[62753]: DEBUG oslo_vmware.api [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': task-1332321, 'name': PowerOnVM_Task, 'duration_secs': 0.652445} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 900.618803] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 900.619015] env[62753]: INFO nova.compute.manager [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Took 8.99 seconds to spawn the instance on the hypervisor. [ 900.619207] env[62753]: DEBUG nova.compute.manager [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 900.620019] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a94ca00-91f4-467e-9ab2-fa4cf5c3a8c7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 900.625743] env[62753]: DEBUG oslo_vmware.api [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332326, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.707757] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332322, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.789607] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332324, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 900.796019] env[62753]: DEBUG nova.scheduler.client.report [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 901.109719] env[62753]: DEBUG nova.network.neutron [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 901.120847] env[62753]: DEBUG oslo_vmware.api [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332326, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.144327] env[62753]: INFO nova.compute.manager [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Took 32.27 seconds to build instance. [ 901.207268] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332322, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.258501] env[62753]: DEBUG nova.network.neutron [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating instance_info_cache with network_info: [{"id": "81591716-fb4a-427f-b3e9-a8b265fe1ce1", "address": "fa:16:3e:3e:ef:2c", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81591716-fb", "ovs_interfaceid": "81591716-fb4a-427f-b3e9-a8b265fe1ce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 901.289571] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332324, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.303029] env[62753]: DEBUG oslo_concurrency.lockutils [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.901s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.303835] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 19.199s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.303835] env[62753]: DEBUG nova.objects.instance [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62753) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 901.332332] env[62753]: INFO nova.scheduler.client.report [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Deleted allocations for instance 046fc3dc-a187-4fda-bc66-345e1226f83d [ 901.621158] env[62753]: DEBUG oslo_vmware.api [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Task: {'id': task-1332326, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.724162} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.621438] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 901.621605] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 901.621779] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 901.621955] env[62753]: INFO nova.compute.manager [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Took 1.64 seconds to destroy the instance on the hypervisor. [ 901.622228] env[62753]: DEBUG oslo.service.loopingcall [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.622741] env[62753]: DEBUG nova.compute.manager [-] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 901.622842] env[62753]: DEBUG nova.network.neutron [-] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 901.642418] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "f03739f4-921f-46e5-b0e7-e9a99d96d621" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.642637] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "f03739f4-921f-46e5-b0e7-e9a99d96d621" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.642875] env[62753]: INFO nova.compute.manager [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Attaching volume f9588f70-4cc6-40d4-9c38-0fda98e95283 to /dev/sdb [ 901.645984] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ae312a85-0a8a-43ad-8fef-70cea18431df tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Lock "9a8a148c-057d-4c91-8b7c-4e05e6ac0f66" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.757s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.674870] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e4e1ce3-fafd-4030-a2e6-37f27000d992 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.683945] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36addeff-7cb9-4013-ab6e-72bb125aa0f3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.699605] env[62753]: DEBUG nova.virt.block_device [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Updating existing volume attachment record: 9adbaf05-ff13-4a93-96f7-831270e27adf {{(pid=62753) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 901.709761] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332322, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.658328} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 901.710085] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 6b9a08b0-3cb3-474a-8193-953e74252a53/6b9a08b0-3cb3-474a-8193-953e74252a53.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 901.710310] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 901.710564] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-74e58b65-f858-4abc-bfdf-5793c521869e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.716781] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 901.716781] env[62753]: value = "task-1332327" [ 901.716781] env[62753]: _type = "Task" [ 901.716781] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.724516] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332327, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.760665] env[62753]: DEBUG oslo_concurrency.lockutils [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 901.760665] env[62753]: DEBUG nova.compute.manager [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Instance network_info: |[{"id": "81591716-fb4a-427f-b3e9-a8b265fe1ce1", "address": "fa:16:3e:3e:ef:2c", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81591716-fb", "ovs_interfaceid": "81591716-fb4a-427f-b3e9-a8b265fe1ce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 901.761207] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3e:ef:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '81591716-fb4a-427f-b3e9-a8b265fe1ce1', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 901.769585] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Creating folder: Project (196457eaa51643c38b4ef4fc6fbbcba3). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 901.770778] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-952b24b2-1d10-4a50-be61-6b5fa02c11bc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.785880] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Created folder: Project (196457eaa51643c38b4ef4fc6fbbcba3) in parent group-v284541. [ 901.786099] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Creating folder: Instances. Parent ref: group-v284608. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 901.786695] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5f257900-84be-45df-a067-fc4b35ea3c82 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.791536] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332324, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.798457] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Created folder: Instances in parent group-v284608. [ 901.798696] env[62753]: DEBUG oslo.service.loopingcall [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 901.798886] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 901.799118] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a97e7e1c-c29c-402e-a2d8-c8a5d810c434 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 901.822425] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 901.822425] env[62753]: value = "task-1332330" [ 901.822425] env[62753]: _type = "Task" [ 901.822425] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 901.831807] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332330, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 901.840091] env[62753]: DEBUG oslo_concurrency.lockutils [None req-508dc3fc-4dda-4a2e-bdf9-2ed07f86ae6c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "046fc3dc-a187-4fda-bc66-345e1226f83d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 28.336s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 901.981979] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "interface-3bc05fdb-81be-4764-b9d9-01acd125b020-43ffe16d-d249-47ae-842c-564f9274c653" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 901.982277] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "interface-3bc05fdb-81be-4764-b9d9-01acd125b020-43ffe16d-d249-47ae-842c-564f9274c653" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 901.982677] env[62753]: DEBUG nova.objects.instance [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lazy-loading 'flavor' on Instance uuid 3bc05fdb-81be-4764-b9d9-01acd125b020 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.226907] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332327, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067374} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.227229] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 902.228515] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-10a35b47-45ff-4858-8ff1-d2d6576c8627 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.251170] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Reconfiguring VM instance instance-00000047 to attach disk [datastore2] 6b9a08b0-3cb3-474a-8193-953e74252a53/6b9a08b0-3cb3-474a-8193-953e74252a53.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 902.251536] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01412f39-6772-41ad-9a79-0472f86fe910 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.266813] env[62753]: DEBUG oslo_concurrency.lockutils [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Acquiring lock "9a8a148c-057d-4c91-8b7c-4e05e6ac0f66" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.267090] env[62753]: DEBUG oslo_concurrency.lockutils [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Lock "9a8a148c-057d-4c91-8b7c-4e05e6ac0f66" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.267376] env[62753]: DEBUG oslo_concurrency.lockutils [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Acquiring lock "9a8a148c-057d-4c91-8b7c-4e05e6ac0f66-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.267598] env[62753]: DEBUG oslo_concurrency.lockutils [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Lock "9a8a148c-057d-4c91-8b7c-4e05e6ac0f66-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.267781] env[62753]: DEBUG oslo_concurrency.lockutils [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Lock "9a8a148c-057d-4c91-8b7c-4e05e6ac0f66-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.272075] env[62753]: INFO nova.compute.manager [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Terminating instance [ 902.274728] env[62753]: DEBUG nova.compute.manager [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 902.275016] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 902.275943] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e4dffaa-1e6f-471c-bb25-b46b85d297c0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.283605] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 902.283605] env[62753]: value = "task-1332334" [ 902.283605] env[62753]: _type = "Task" [ 902.283605] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.295705] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 902.296389] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4a1a4763-47a7-426e-b8c6-63f673a5770c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.305451] env[62753]: DEBUG oslo_vmware.api [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332324, 'name': PowerOnVM_Task, 'duration_secs': 2.020455} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.305709] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332334, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.306402] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 902.306597] env[62753]: DEBUG nova.compute.manager [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 902.307391] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95be1fa5-ffba-4274-a8c9-3ff07c3cbab3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.312019] env[62753]: DEBUG oslo_vmware.api [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Waiting for the task: (returnval){ [ 902.312019] env[62753]: value = "task-1332335" [ 902.312019] env[62753]: _type = "Task" [ 902.312019] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.321323] env[62753]: DEBUG oslo_concurrency.lockutils [None req-36beaaf4-e564-4f59-87bf-80124a3c0b84 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.017s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.322730] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 16.804s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.322730] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 902.322730] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62753) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 902.322730] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 16.016s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 902.322730] env[62753]: DEBUG nova.objects.instance [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lazy-loading 'resources' on Instance uuid a6ee010c-6305-4009-80e0-92a2c58bcd7b {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.324831] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06d94c1a-a0ca-4cc1-bac6-55980522bf82 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.334917] env[62753]: DEBUG oslo_vmware.api [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': task-1332335, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.344550] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b6a31ef-08af-404b-92b5-f725655be85e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.349248] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332330, 'name': CreateVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.362652] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-49e26477-16d9-4617-8140-5c1224bd7ff8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.371887] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c06f15c-0255-4842-bf89-2b89fb9a1fc7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.405070] env[62753]: DEBUG nova.network.neutron [-] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 902.406387] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181310MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=62753) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 902.406542] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.597313] env[62753]: DEBUG nova.objects.instance [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lazy-loading 'pci_requests' on Instance uuid 3bc05fdb-81be-4764-b9d9-01acd125b020 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 902.603919] env[62753]: DEBUG nova.compute.manager [req-1d465b23-5515-4c1a-9081-9e824d48294f req-6ec49b54-c635-4f5a-8201-4ee40fd777c7 service nova] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Received event network-changed-81591716-fb4a-427f-b3e9-a8b265fe1ce1 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 902.604241] env[62753]: DEBUG nova.compute.manager [req-1d465b23-5515-4c1a-9081-9e824d48294f req-6ec49b54-c635-4f5a-8201-4ee40fd777c7 service nova] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Refreshing instance network info cache due to event network-changed-81591716-fb4a-427f-b3e9-a8b265fe1ce1. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 902.604486] env[62753]: DEBUG oslo_concurrency.lockutils [req-1d465b23-5515-4c1a-9081-9e824d48294f req-6ec49b54-c635-4f5a-8201-4ee40fd777c7 service nova] Acquiring lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.604528] env[62753]: DEBUG oslo_concurrency.lockutils [req-1d465b23-5515-4c1a-9081-9e824d48294f req-6ec49b54-c635-4f5a-8201-4ee40fd777c7 service nova] Acquired lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.604707] env[62753]: DEBUG nova.network.neutron [req-1d465b23-5515-4c1a-9081-9e824d48294f req-6ec49b54-c635-4f5a-8201-4ee40fd777c7 service nova] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Refreshing network info cache for port 81591716-fb4a-427f-b3e9-a8b265fe1ce1 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 902.799146] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332334, 'name': ReconfigVM_Task, 'duration_secs': 0.335217} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.799479] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Reconfigured VM instance instance-00000047 to attach disk [datastore2] 6b9a08b0-3cb3-474a-8193-953e74252a53/6b9a08b0-3cb3-474a-8193-953e74252a53.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 902.800127] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-39d0cab1-f4a3-4191-9040-6587da902d03 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.807305] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 902.807305] env[62753]: value = "task-1332336" [ 902.807305] env[62753]: _type = "Task" [ 902.807305] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.815657] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332336, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.825527] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 902.831180] env[62753]: DEBUG oslo_vmware.api [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': task-1332335, 'name': PowerOffVM_Task, 'duration_secs': 0.251015} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.835345] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 902.835647] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 902.835977] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-936c2df2-e844-4f4b-9254-2dfec8215419 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.844596] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332330, 'name': CreateVM_Task, 'duration_secs': 0.52693} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 902.844777] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 902.845494] env[62753]: DEBUG oslo_concurrency.lockutils [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 902.845703] env[62753]: DEBUG oslo_concurrency.lockutils [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 902.846097] env[62753]: DEBUG oslo_concurrency.lockutils [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 902.846399] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-389b585e-54df-43d5-b32c-087e0f571b31 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.853143] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 902.853143] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f6f995-818f-aad0-d49a-2de6c6e0f8cb" [ 902.853143] env[62753]: _type = "Task" [ 902.853143] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.861774] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f6f995-818f-aad0-d49a-2de6c6e0f8cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 902.908846] env[62753]: INFO nova.compute.manager [-] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Took 1.29 seconds to deallocate network for instance. [ 902.909239] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 902.909483] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 902.909762] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Deleting the datastore file [datastore1] 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 902.913935] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0dd32c44-4432-4643-b698-2a04c946a395 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 902.925205] env[62753]: DEBUG oslo_vmware.api [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Waiting for the task: (returnval){ [ 902.925205] env[62753]: value = "task-1332338" [ 902.925205] env[62753]: _type = "Task" [ 902.925205] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 902.938609] env[62753]: DEBUG oslo_vmware.api [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': task-1332338, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.053795] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b4cf0f6-e9d3-4368-afcd-baf70aca9a88 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.061836] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c1493a-1e4f-4230-931f-a5903f446781 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.091223] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75792e4f-96bc-479a-93be-9412346e3b99 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.100197] env[62753]: DEBUG nova.objects.base [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Object Instance<3bc05fdb-81be-4764-b9d9-01acd125b020> lazy-loaded attributes: flavor,pci_requests {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 903.100510] env[62753]: DEBUG nova.network.neutron [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 903.104082] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2fb16f2-64f5-4230-a184-9e765c4b6a9e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.123752] env[62753]: DEBUG nova.compute.provider_tree [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 903.174084] env[62753]: DEBUG nova.policy [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9931b40cfd7846038805c6a4caedaac0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b8e81660b30408c998e412f5fa81469', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 903.317440] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332336, 'name': Rename_Task, 'duration_secs': 0.14814} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.317724] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 903.317988] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7653445b-bc22-463c-abd6-285a0a09793d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.324841] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 903.324841] env[62753]: value = "task-1332339" [ 903.324841] env[62753]: _type = "Task" [ 903.324841] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.332658] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332339, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.364072] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f6f995-818f-aad0-d49a-2de6c6e0f8cb, 'name': SearchDatastore_Task, 'duration_secs': 0.010824} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.364182] env[62753]: DEBUG oslo_concurrency.lockutils [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.365045] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 903.365045] env[62753]: DEBUG oslo_concurrency.lockutils [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 903.365045] env[62753]: DEBUG oslo_concurrency.lockutils [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 903.365045] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 903.365541] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0f2c5121-369c-445d-be9a-8b4b0c0cd7c0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.374761] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 903.375151] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 903.376038] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3ef2bb0-134e-4ba1-b972-7947a2571c82 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.383626] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 903.383626] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f154fc-d243-6cdf-0f6a-14a3519fbe51" [ 903.383626] env[62753]: _type = "Task" [ 903.383626] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.392464] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f154fc-d243-6cdf-0f6a-14a3519fbe51, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.403364] env[62753]: DEBUG nova.network.neutron [req-1d465b23-5515-4c1a-9081-9e824d48294f req-6ec49b54-c635-4f5a-8201-4ee40fd777c7 service nova] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updated VIF entry in instance network info cache for port 81591716-fb4a-427f-b3e9-a8b265fe1ce1. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 903.404055] env[62753]: DEBUG nova.network.neutron [req-1d465b23-5515-4c1a-9081-9e824d48294f req-6ec49b54-c635-4f5a-8201-4ee40fd777c7 service nova] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating instance_info_cache with network_info: [{"id": "81591716-fb4a-427f-b3e9-a8b265fe1ce1", "address": "fa:16:3e:3e:ef:2c", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81591716-fb", "ovs_interfaceid": "81591716-fb4a-427f-b3e9-a8b265fe1ce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.421338] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 903.435517] env[62753]: DEBUG oslo_vmware.api [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Task: {'id': task-1332338, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.397699} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.435764] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 903.435952] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 903.436151] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 903.436326] env[62753]: INFO nova.compute.manager [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Took 1.16 seconds to destroy the instance on the hypervisor. [ 903.436570] env[62753]: DEBUG oslo.service.loopingcall [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 903.436759] env[62753]: DEBUG nova.compute.manager [-] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 903.436855] env[62753]: DEBUG nova.network.neutron [-] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 903.627490] env[62753]: DEBUG nova.scheduler.client.report [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 903.778091] env[62753]: DEBUG nova.compute.manager [req-534839a8-8956-4585-87c7-2da685a4f3b7 req-5d4cb254-771e-45e3-82a0-110951814997 service nova] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Received event network-vif-deleted-61fe245f-aaa8-4566-857d-3002a5c3426d {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.778307] env[62753]: INFO nova.compute.manager [req-534839a8-8956-4585-87c7-2da685a4f3b7 req-5d4cb254-771e-45e3-82a0-110951814997 service nova] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Neutron deleted interface 61fe245f-aaa8-4566-857d-3002a5c3426d; detaching it from the instance and deleting it from the info cache [ 903.778489] env[62753]: DEBUG nova.network.neutron [req-534839a8-8956-4585-87c7-2da685a4f3b7 req-5d4cb254-771e-45e3-82a0-110951814997 service nova] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 903.836088] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332339, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 903.895770] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f154fc-d243-6cdf-0f6a-14a3519fbe51, 'name': SearchDatastore_Task, 'duration_secs': 0.027043} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.897249] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d12e1a7d-e03c-4166-9fda-2123aaba5d82 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.907275] env[62753]: DEBUG oslo_concurrency.lockutils [req-1d465b23-5515-4c1a-9081-9e824d48294f req-6ec49b54-c635-4f5a-8201-4ee40fd777c7 service nova] Releasing lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.907599] env[62753]: DEBUG nova.compute.manager [req-1d465b23-5515-4c1a-9081-9e824d48294f req-6ec49b54-c635-4f5a-8201-4ee40fd777c7 service nova] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Received event network-vif-deleted-1f18d3a5-8537-41a7-adcc-ea26fe748a8e {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 903.908473] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 903.908473] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5288b3c5-5a76-ed6a-6be3-1cbe84c7b8b6" [ 903.908473] env[62753]: _type = "Task" [ 903.908473] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.921106] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5288b3c5-5a76-ed6a-6be3-1cbe84c7b8b6, 'name': SearchDatastore_Task, 'duration_secs': 0.011941} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 903.921464] env[62753]: DEBUG oslo_concurrency.lockutils [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 903.923643] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] c73b1ae0-dc98-47f7-babf-e96169384785/c73b1ae0-dc98-47f7-babf-e96169384785.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 903.925386] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7a02abe5-1735-48bd-a53f-2c389b01dfb0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 903.933078] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 903.933078] env[62753]: value = "task-1332340" [ 903.933078] env[62753]: _type = "Task" [ 903.933078] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 903.942992] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332340, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.133341] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.811s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.137047] env[62753]: DEBUG oslo_concurrency.lockutils [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 15.181s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.139388] env[62753]: INFO nova.compute.claims [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 904.162398] env[62753]: INFO nova.scheduler.client.report [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Deleted allocations for instance a6ee010c-6305-4009-80e0-92a2c58bcd7b [ 904.168632] env[62753]: DEBUG nova.network.neutron [-] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.281893] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9bd0b279-4d0e-457d-a909-47543b6f0597 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.294832] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bdad6d6-0888-4e83-bca3-9b2c88011546 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.326518] env[62753]: DEBUG nova.compute.manager [req-534839a8-8956-4585-87c7-2da685a4f3b7 req-5d4cb254-771e-45e3-82a0-110951814997 service nova] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Detach interface failed, port_id=61fe245f-aaa8-4566-857d-3002a5c3426d, reason: Instance 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 904.335775] env[62753]: DEBUG oslo_vmware.api [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332339, 'name': PowerOnVM_Task, 'duration_secs': 0.828111} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.336087] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 904.336320] env[62753]: INFO nova.compute.manager [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Took 9.20 seconds to spawn the instance on the hypervisor. [ 904.336509] env[62753]: DEBUG nova.compute.manager [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 904.337338] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed864788-746f-4710-ac0c-cfbf818611ce {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.373658] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "33c1de85-aade-4c69-8fdb-7672d8c53300" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.374053] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "33c1de85-aade-4c69-8fdb-7672d8c53300" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.374294] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "33c1de85-aade-4c69-8fdb-7672d8c53300-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.374962] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "33c1de85-aade-4c69-8fdb-7672d8c53300-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.375185] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "33c1de85-aade-4c69-8fdb-7672d8c53300-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.377996] env[62753]: INFO nova.compute.manager [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Terminating instance [ 904.379984] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "refresh_cache-33c1de85-aade-4c69-8fdb-7672d8c53300" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 904.380170] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquired lock "refresh_cache-33c1de85-aade-4c69-8fdb-7672d8c53300" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 904.380347] env[62753]: DEBUG nova.network.neutron [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 904.444495] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332340, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 904.632417] env[62753]: DEBUG nova.compute.manager [req-25b84825-448f-4e28-9f1d-9c3e52641f32 req-7feb17cf-414d-47d3-8468-2e2f4a001b65 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Received event network-vif-plugged-43ffe16d-d249-47ae-842c-564f9274c653 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 904.632417] env[62753]: DEBUG oslo_concurrency.lockutils [req-25b84825-448f-4e28-9f1d-9c3e52641f32 req-7feb17cf-414d-47d3-8468-2e2f4a001b65 service nova] Acquiring lock "3bc05fdb-81be-4764-b9d9-01acd125b020-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 904.632417] env[62753]: DEBUG oslo_concurrency.lockutils [req-25b84825-448f-4e28-9f1d-9c3e52641f32 req-7feb17cf-414d-47d3-8468-2e2f4a001b65 service nova] Lock "3bc05fdb-81be-4764-b9d9-01acd125b020-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 904.632417] env[62753]: DEBUG oslo_concurrency.lockutils [req-25b84825-448f-4e28-9f1d-9c3e52641f32 req-7feb17cf-414d-47d3-8468-2e2f4a001b65 service nova] Lock "3bc05fdb-81be-4764-b9d9-01acd125b020-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.632417] env[62753]: DEBUG nova.compute.manager [req-25b84825-448f-4e28-9f1d-9c3e52641f32 req-7feb17cf-414d-47d3-8468-2e2f4a001b65 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] No waiting events found dispatching network-vif-plugged-43ffe16d-d249-47ae-842c-564f9274c653 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 904.632842] env[62753]: WARNING nova.compute.manager [req-25b84825-448f-4e28-9f1d-9c3e52641f32 req-7feb17cf-414d-47d3-8468-2e2f4a001b65 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Received unexpected event network-vif-plugged-43ffe16d-d249-47ae-842c-564f9274c653 for instance with vm_state active and task_state None. [ 904.673026] env[62753]: INFO nova.compute.manager [-] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Took 1.24 seconds to deallocate network for instance. [ 904.673026] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5acd5fa5-e7ab-4a78-8486-018db10649b9 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "a6ee010c-6305-4009-80e0-92a2c58bcd7b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.577s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 904.718647] env[62753]: DEBUG nova.network.neutron [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Successfully updated port: 43ffe16d-d249-47ae-842c-564f9274c653 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 904.854543] env[62753]: INFO nova.compute.manager [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Took 32.99 seconds to build instance. [ 904.898041] env[62753]: DEBUG nova.network.neutron [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 904.945220] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332340, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.582031} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 904.945511] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] c73b1ae0-dc98-47f7-babf-e96169384785/c73b1ae0-dc98-47f7-babf-e96169384785.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 904.945734] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 904.945992] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-6c94d365-094a-499f-a806-2363d406c41e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 904.953618] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 904.953618] env[62753]: value = "task-1332342" [ 904.953618] env[62753]: _type = "Task" [ 904.953618] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 904.957497] env[62753]: DEBUG nova.network.neutron [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 904.964129] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332342, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.179476] env[62753]: DEBUG oslo_concurrency.lockutils [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 905.221875] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 905.222104] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 905.222331] env[62753]: DEBUG nova.network.neutron [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 905.358760] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a64253c-26a9-46d8-b561-ea310e6c2ca7 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "6b9a08b0-3cb3-474a-8193-953e74252a53" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 57.290s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 905.359818] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3dd18d92-9ce0-48c2-ab05-341241f8b49b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.368223] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b66a37a8-f178-4d36-a192-63c0906c8e64 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.399018] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819174f2-d334-4a10-b563-02abc8b38dbf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.409440] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1e86f3d-4fac-407b-b407-cc5ee9185e35 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.428226] env[62753]: DEBUG nova.compute.provider_tree [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 905.460239] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Releasing lock "refresh_cache-33c1de85-aade-4c69-8fdb-7672d8c53300" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 905.460662] env[62753]: DEBUG nova.compute.manager [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 905.460860] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 905.464773] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f1ec979-79d3-4b07-9eb4-6972edd319d3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.467396] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332342, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06909} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 905.467637] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 905.468734] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c32636bd-25ee-46ea-a29b-2704425c8354 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.473506] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 905.474062] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-850bcf1c-78fe-47eb-bca4-79e1a9d5f507 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.494734] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] c73b1ae0-dc98-47f7-babf-e96169384785/c73b1ae0-dc98-47f7-babf-e96169384785.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 905.496440] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-01930434-4f34-47ad-91e7-7371c07d30b1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 905.510921] env[62753]: DEBUG oslo_vmware.api [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 905.510921] env[62753]: value = "task-1332343" [ 905.510921] env[62753]: _type = "Task" [ 905.510921] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.518553] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 905.518553] env[62753]: value = "task-1332344" [ 905.518553] env[62753]: _type = "Task" [ 905.518553] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 905.521729] env[62753]: DEBUG oslo_vmware.api [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332343, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.529873] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332344, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 905.760920] env[62753]: WARNING nova.network.neutron [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] 8a9463a1-a861-4ade-8483-640392be0a89 already exists in list: networks containing: ['8a9463a1-a861-4ade-8483-640392be0a89']. ignoring it [ 905.761190] env[62753]: WARNING nova.network.neutron [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] 8a9463a1-a861-4ade-8483-640392be0a89 already exists in list: networks containing: ['8a9463a1-a861-4ade-8483-640392be0a89']. ignoring it [ 905.936683] env[62753]: DEBUG nova.scheduler.client.report [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 906.022802] env[62753]: DEBUG oslo_vmware.api [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332343, 'name': PowerOffVM_Task, 'duration_secs': 0.14612} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.026276] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 906.026492] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 906.027317] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3adb5e2c-ba7a-415c-b653-8d24385e50f8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.034937] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332344, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.055127] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 906.055416] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 906.055550] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Deleting the datastore file [datastore2] 33c1de85-aade-4c69-8fdb-7672d8c53300 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.055832] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2bb86ae6-e268-4243-8725-83f78696e276 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.062377] env[62753]: DEBUG oslo_vmware.api [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 906.062377] env[62753]: value = "task-1332346" [ 906.062377] env[62753]: _type = "Task" [ 906.062377] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.071076] env[62753]: DEBUG oslo_vmware.api [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332346, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.217709] env[62753]: DEBUG nova.network.neutron [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Updating instance_info_cache with network_info: [{"id": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "address": "fa:16:3e:1d:84:1b", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8e502a-7a", "ovs_interfaceid": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9713beba-767d-408c-bd7d-bef58dc3c405", "address": "fa:16:3e:da:85:43", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9713beba-76", "ovs_interfaceid": "9713beba-767d-408c-bd7d-bef58dc3c405", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "43ffe16d-d249-47ae-842c-564f9274c653", "address": "fa:16:3e:a8:c4:dc", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43ffe16d-d2", "ovs_interfaceid": "43ffe16d-d249-47ae-842c-564f9274c653", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 906.256168] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Volume attach. Driver type: vmdk {{(pid=62753) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 906.256478] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284612', 'volume_id': 'f9588f70-4cc6-40d4-9c38-0fda98e95283', 'name': 'volume-f9588f70-4cc6-40d4-9c38-0fda98e95283', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f03739f4-921f-46e5-b0e7-e9a99d96d621', 'attached_at': '', 'detached_at': '', 'volume_id': 'f9588f70-4cc6-40d4-9c38-0fda98e95283', 'serial': 'f9588f70-4cc6-40d4-9c38-0fda98e95283'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 906.257450] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72415d87-7043-4c7a-806d-6b18c7ab9f1f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.276727] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f729169-f74a-4ef5-95d6-5541db4e9a46 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.301579] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Reconfiguring VM instance instance-00000045 to attach disk [datastore1] volume-f9588f70-4cc6-40d4-9c38-0fda98e95283/volume-f9588f70-4cc6-40d4-9c38-0fda98e95283.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 906.301908] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4a92d9c6-64f2-4baf-8e70-a03672c26f4d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.320353] env[62753]: DEBUG oslo_vmware.api [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 906.320353] env[62753]: value = "task-1332347" [ 906.320353] env[62753]: _type = "Task" [ 906.320353] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.328964] env[62753]: DEBUG oslo_vmware.api [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332347, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.346722] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "6b9a08b0-3cb3-474a-8193-953e74252a53" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.347031] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "6b9a08b0-3cb3-474a-8193-953e74252a53" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.347253] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "6b9a08b0-3cb3-474a-8193-953e74252a53-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 906.347453] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "6b9a08b0-3cb3-474a-8193-953e74252a53-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.347651] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "6b9a08b0-3cb3-474a-8193-953e74252a53-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.350194] env[62753]: INFO nova.compute.manager [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Terminating instance [ 906.352459] env[62753]: DEBUG nova.compute.manager [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 906.352708] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 906.353922] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a2c9884-916f-4995-a36a-e47431294922 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.364073] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 906.364073] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-edb72ced-12d5-4683-8bef-15a0c6174de1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.369445] env[62753]: DEBUG oslo_vmware.api [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 906.369445] env[62753]: value = "task-1332348" [ 906.369445] env[62753]: _type = "Task" [ 906.369445] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.377803] env[62753]: DEBUG oslo_vmware.api [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332348, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.444692] env[62753]: DEBUG oslo_concurrency.lockutils [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.308s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 906.445241] env[62753]: DEBUG nova.compute.manager [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 906.447955] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 16.808s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 906.449455] env[62753]: INFO nova.compute.claims [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 906.531582] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332344, 'name': ReconfigVM_Task, 'duration_secs': 0.797913} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.531956] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Reconfigured VM instance instance-00000048 to attach disk [datastore1] c73b1ae0-dc98-47f7-babf-e96169384785/c73b1ae0-dc98-47f7-babf-e96169384785.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.532638] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ec93939a-0bcf-45f0-bf70-59c58091ec34 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.538973] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 906.538973] env[62753]: value = "task-1332349" [ 906.538973] env[62753]: _type = "Task" [ 906.538973] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.547536] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332349, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.573049] env[62753]: DEBUG oslo_vmware.api [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332346, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.268998} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.573386] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 906.573620] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 906.573863] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 906.574099] env[62753]: INFO nova.compute.manager [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Took 1.11 seconds to destroy the instance on the hypervisor. [ 906.574416] env[62753]: DEBUG oslo.service.loopingcall [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 906.574696] env[62753]: DEBUG nova.compute.manager [-] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 906.574822] env[62753]: DEBUG nova.network.neutron [-] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 906.592928] env[62753]: DEBUG nova.network.neutron [-] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 906.658854] env[62753]: DEBUG nova.compute.manager [req-3d8d715c-eaef-4ef4-992b-ae8544e905f8 req-a9119e05-62f9-43d8-ae1d-475b06117a9a service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Received event network-changed-43ffe16d-d249-47ae-842c-564f9274c653 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 906.658854] env[62753]: DEBUG nova.compute.manager [req-3d8d715c-eaef-4ef4-992b-ae8544e905f8 req-a9119e05-62f9-43d8-ae1d-475b06117a9a service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Refreshing instance network info cache due to event network-changed-43ffe16d-d249-47ae-842c-564f9274c653. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 906.659106] env[62753]: DEBUG oslo_concurrency.lockutils [req-3d8d715c-eaef-4ef4-992b-ae8544e905f8 req-a9119e05-62f9-43d8-ae1d-475b06117a9a service nova] Acquiring lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.720649] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 906.721378] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 906.721528] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.721817] env[62753]: DEBUG oslo_concurrency.lockutils [req-3d8d715c-eaef-4ef4-992b-ae8544e905f8 req-a9119e05-62f9-43d8-ae1d-475b06117a9a service nova] Acquired lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 906.721998] env[62753]: DEBUG nova.network.neutron [req-3d8d715c-eaef-4ef4-992b-ae8544e905f8 req-a9119e05-62f9-43d8-ae1d-475b06117a9a service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Refreshing network info cache for port 43ffe16d-d249-47ae-842c-564f9274c653 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 906.723920] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73422bf-3aa5-456e-9f93-e319bfb86c9d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.743941] env[62753]: DEBUG nova.virt.hardware [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 906.744243] env[62753]: DEBUG nova.virt.hardware [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 906.744397] env[62753]: DEBUG nova.virt.hardware [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 906.744586] env[62753]: DEBUG nova.virt.hardware [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 906.744735] env[62753]: DEBUG nova.virt.hardware [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 906.744884] env[62753]: DEBUG nova.virt.hardware [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 906.745108] env[62753]: DEBUG nova.virt.hardware [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 906.745306] env[62753]: DEBUG nova.virt.hardware [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 906.745467] env[62753]: DEBUG nova.virt.hardware [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 906.745635] env[62753]: DEBUG nova.virt.hardware [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 906.745814] env[62753]: DEBUG nova.virt.hardware [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 906.752130] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Reconfiguring VM to attach interface {{(pid=62753) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 906.752733] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6b4a5f24-4892-4970-bc81-317485402d2c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.769441] env[62753]: DEBUG oslo_vmware.api [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 906.769441] env[62753]: value = "task-1332350" [ 906.769441] env[62753]: _type = "Task" [ 906.769441] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.777051] env[62753]: DEBUG oslo_vmware.api [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332350, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.830747] env[62753]: DEBUG oslo_vmware.api [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332347, 'name': ReconfigVM_Task, 'duration_secs': 0.424075} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.831059] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Reconfigured VM instance instance-00000045 to attach disk [datastore1] volume-f9588f70-4cc6-40d4-9c38-0fda98e95283/volume-f9588f70-4cc6-40d4-9c38-0fda98e95283.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 906.835893] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-097f52fa-3d2a-41ad-97f8-d79d024d64f0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.851691] env[62753]: DEBUG oslo_vmware.api [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 906.851691] env[62753]: value = "task-1332351" [ 906.851691] env[62753]: _type = "Task" [ 906.851691] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.859543] env[62753]: DEBUG oslo_vmware.api [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332351, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 906.879580] env[62753]: DEBUG oslo_vmware.api [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332348, 'name': PowerOffVM_Task, 'duration_secs': 0.224457} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 906.879896] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 906.880082] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 906.880346] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0c30e9f9-aea3-4766-8f85-c29ff986d9f9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.954098] env[62753]: DEBUG nova.compute.utils [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 906.959435] env[62753]: DEBUG nova.compute.manager [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 906.959668] env[62753]: DEBUG nova.network.neutron [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 906.962043] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 906.962275] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 906.962462] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Deleting the datastore file [datastore2] 6b9a08b0-3cb3-474a-8193-953e74252a53 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 906.962945] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-103a3250-a912-44fa-ac93-d8c79fa75c0d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 906.971119] env[62753]: DEBUG oslo_vmware.api [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 906.971119] env[62753]: value = "task-1332353" [ 906.971119] env[62753]: _type = "Task" [ 906.971119] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 906.979456] env[62753]: DEBUG oslo_vmware.api [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332353, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.003021] env[62753]: DEBUG nova.policy [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '14bfb2a920864f73bfb5fc7a45e7bde8', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '909db6176c8f475caf2562a593bc61a9', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 907.048527] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332349, 'name': Rename_Task, 'duration_secs': 0.155812} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.048825] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 907.049094] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6ab90ef7-f987-48c9-ae51-25704d8af5b0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.055327] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 907.055327] env[62753]: value = "task-1332354" [ 907.055327] env[62753]: _type = "Task" [ 907.055327] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 907.063230] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332354, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.095807] env[62753]: DEBUG nova.network.neutron [-] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.282518] env[62753]: DEBUG oslo_vmware.api [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332350, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.293539] env[62753]: DEBUG nova.network.neutron [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Successfully created port: 4037b488-2f26-40ee-b044-3b4fec79a454 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 907.360823] env[62753]: DEBUG oslo_vmware.api [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332351, 'name': ReconfigVM_Task, 'duration_secs': 0.146096} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.361274] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284612', 'volume_id': 'f9588f70-4cc6-40d4-9c38-0fda98e95283', 'name': 'volume-f9588f70-4cc6-40d4-9c38-0fda98e95283', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f03739f4-921f-46e5-b0e7-e9a99d96d621', 'attached_at': '', 'detached_at': '', 'volume_id': 'f9588f70-4cc6-40d4-9c38-0fda98e95283', 'serial': 'f9588f70-4cc6-40d4-9c38-0fda98e95283'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 907.461100] env[62753]: DEBUG nova.compute.manager [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 907.486951] env[62753]: DEBUG oslo_vmware.api [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332353, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.243295} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.487348] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 907.487548] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 907.487796] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 907.488020] env[62753]: INFO nova.compute.manager [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Took 1.14 seconds to destroy the instance on the hypervisor. [ 907.488284] env[62753]: DEBUG oslo.service.loopingcall [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 907.488488] env[62753]: DEBUG nova.compute.manager [-] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 907.488584] env[62753]: DEBUG nova.network.neutron [-] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 907.521918] env[62753]: DEBUG nova.network.neutron [req-3d8d715c-eaef-4ef4-992b-ae8544e905f8 req-a9119e05-62f9-43d8-ae1d-475b06117a9a service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Updated VIF entry in instance network info cache for port 43ffe16d-d249-47ae-842c-564f9274c653. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 907.522411] env[62753]: DEBUG nova.network.neutron [req-3d8d715c-eaef-4ef4-992b-ae8544e905f8 req-a9119e05-62f9-43d8-ae1d-475b06117a9a service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Updating instance_info_cache with network_info: [{"id": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "address": "fa:16:3e:1d:84:1b", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8e502a-7a", "ovs_interfaceid": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "9713beba-767d-408c-bd7d-bef58dc3c405", "address": "fa:16:3e:da:85:43", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9713beba-76", "ovs_interfaceid": "9713beba-767d-408c-bd7d-bef58dc3c405", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "43ffe16d-d249-47ae-842c-564f9274c653", "address": "fa:16:3e:a8:c4:dc", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43ffe16d-d2", "ovs_interfaceid": "43ffe16d-d249-47ae-842c-564f9274c653", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 907.566470] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332354, 'name': PowerOnVM_Task} progress is 78%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 907.598070] env[62753]: INFO nova.compute.manager [-] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Took 1.02 seconds to deallocate network for instance. [ 907.698207] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ea73484-66cf-4f32-a146-356d3a1f1e1a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.706321] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb879c67-0c14-41d4-a41f-89151561bd74 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.739202] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84ed13bf-2ff9-4b47-b317-e3064982fcb4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.746739] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1f66e66-b81d-4aa4-a571-f8b73506370b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 907.762266] env[62753]: DEBUG nova.compute.provider_tree [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 907.783283] env[62753]: DEBUG oslo_vmware.api [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332350, 'name': ReconfigVM_Task, 'duration_secs': 0.557862} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 907.783824] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 907.784055] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Reconfigured VM to attach interface {{(pid=62753) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 908.028444] env[62753]: DEBUG oslo_concurrency.lockutils [req-3d8d715c-eaef-4ef4-992b-ae8544e905f8 req-a9119e05-62f9-43d8-ae1d-475b06117a9a service nova] Releasing lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 908.066769] env[62753]: DEBUG oslo_vmware.api [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332354, 'name': PowerOnVM_Task, 'duration_secs': 0.767969} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 908.067763] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 908.067763] env[62753]: INFO nova.compute.manager [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Took 8.60 seconds to spawn the instance on the hypervisor. [ 908.067763] env[62753]: DEBUG nova.compute.manager [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 908.068541] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c500ba74-9c58-40ab-98cc-30617b9e3b01 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.105264] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 908.266773] env[62753]: DEBUG nova.scheduler.client.report [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 908.283932] env[62753]: DEBUG nova.network.neutron [-] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 908.288786] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1216db27-a461-47a0-878d-ba611b91a857 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "interface-3bc05fdb-81be-4764-b9d9-01acd125b020-43ffe16d-d249-47ae-842c-564f9274c653" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.306s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.403063] env[62753]: DEBUG nova.objects.instance [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lazy-loading 'flavor' on Instance uuid f03739f4-921f-46e5-b0e7-e9a99d96d621 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 908.474596] env[62753]: DEBUG nova.compute.manager [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 908.501804] env[62753]: DEBUG nova.virt.hardware [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 908.502064] env[62753]: DEBUG nova.virt.hardware [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 908.502233] env[62753]: DEBUG nova.virt.hardware [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 908.502702] env[62753]: DEBUG nova.virt.hardware [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 908.502702] env[62753]: DEBUG nova.virt.hardware [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 908.502702] env[62753]: DEBUG nova.virt.hardware [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 908.502912] env[62753]: DEBUG nova.virt.hardware [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 908.503737] env[62753]: DEBUG nova.virt.hardware [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 908.504019] env[62753]: DEBUG nova.virt.hardware [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 908.504217] env[62753]: DEBUG nova.virt.hardware [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 908.504411] env[62753]: DEBUG nova.virt.hardware [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 908.505300] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53364621-fe2c-4235-af97-04048e3d7fe9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.513221] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ef1d2b7-1b5a-44b4-99ba-da11a8f36830 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 908.589101] env[62753]: INFO nova.compute.manager [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Took 33.22 seconds to build instance. [ 908.772430] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.324s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.774491] env[62753]: DEBUG nova.compute.manager [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 908.775658] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 18.450s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 908.777032] env[62753]: INFO nova.compute.claims [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 908.786883] env[62753]: INFO nova.compute.manager [-] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Took 1.30 seconds to deallocate network for instance. [ 908.850718] env[62753]: DEBUG nova.network.neutron [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Successfully updated port: 4037b488-2f26-40ee-b044-3b4fec79a454 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 908.907697] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1a155711-4c20-487f-8b8c-839a7ba5faf5 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "f03739f4-921f-46e5-b0e7-e9a99d96d621" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.265s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 908.959569] env[62753]: DEBUG nova.compute.manager [req-07546617-512c-47c1-8510-fb9c7722d75a req-bc2b7efb-34bc-4f58-8caf-69a7c6eb8ba9 service nova] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Received event network-vif-deleted-66ec9e7b-4a31-4ee0-9ae0-f1ecf62c57bf {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.088707] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "f03739f4-921f-46e5-b0e7-e9a99d96d621" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.089081] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "f03739f4-921f-46e5-b0e7-e9a99d96d621" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.089208] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "f03739f4-921f-46e5-b0e7-e9a99d96d621-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.089399] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "f03739f4-921f-46e5-b0e7-e9a99d96d621-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 909.089569] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "f03739f4-921f-46e5-b0e7-e9a99d96d621-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.091847] env[62753]: DEBUG oslo_concurrency.lockutils [None req-279c8503-aec6-4b06-93bd-a5a6d5f0a1c0 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "c73b1ae0-dc98-47f7-babf-e96169384785" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.641s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 909.092358] env[62753]: INFO nova.compute.manager [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Terminating instance [ 909.094446] env[62753]: DEBUG nova.compute.manager [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 909.094771] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 909.095049] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-451efd30-39ef-4362-9ba5-da541953f12d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.102088] env[62753]: DEBUG oslo_vmware.api [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 909.102088] env[62753]: value = "task-1332355" [ 909.102088] env[62753]: _type = "Task" [ 909.102088] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.112163] env[62753]: DEBUG oslo_vmware.api [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332355, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.281971] env[62753]: DEBUG nova.compute.utils [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 909.285546] env[62753]: DEBUG nova.compute.manager [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 909.285731] env[62753]: DEBUG nova.network.neutron [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 909.294130] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 909.330935] env[62753]: DEBUG nova.policy [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '76de60499b104b23bf4a469f70db8d6c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'c2dc6fe36ae84435a8e09cdf56340c58', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 909.353404] env[62753]: DEBUG oslo_concurrency.lockutils [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "refresh_cache-77cdd901-cc96-4a6f-a696-a65f54a96b1a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.353794] env[62753]: DEBUG oslo_concurrency.lockutils [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquired lock "refresh_cache-77cdd901-cc96-4a6f-a696-a65f54a96b1a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.354113] env[62753]: DEBUG nova.network.neutron [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 909.600539] env[62753]: DEBUG nova.network.neutron [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Successfully created port: ef993aa6-10f5-4de9-b652-426dbaac5112 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 909.609444] env[62753]: DEBUG nova.compute.manager [req-1be711d0-4dc1-4674-b223-4ecca1af663a req-5d85b151-6c29-4fcf-92d2-ce5031ce1ad7 service nova] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Received event network-changed-81591716-fb4a-427f-b3e9-a8b265fe1ce1 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 909.609808] env[62753]: DEBUG nova.compute.manager [req-1be711d0-4dc1-4674-b223-4ecca1af663a req-5d85b151-6c29-4fcf-92d2-ce5031ce1ad7 service nova] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Refreshing instance network info cache due to event network-changed-81591716-fb4a-427f-b3e9-a8b265fe1ce1. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 909.610087] env[62753]: DEBUG oslo_concurrency.lockutils [req-1be711d0-4dc1-4674-b223-4ecca1af663a req-5d85b151-6c29-4fcf-92d2-ce5031ce1ad7 service nova] Acquiring lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 909.610430] env[62753]: DEBUG oslo_concurrency.lockutils [req-1be711d0-4dc1-4674-b223-4ecca1af663a req-5d85b151-6c29-4fcf-92d2-ce5031ce1ad7 service nova] Acquired lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 909.610551] env[62753]: DEBUG nova.network.neutron [req-1be711d0-4dc1-4674-b223-4ecca1af663a req-5d85b151-6c29-4fcf-92d2-ce5031ce1ad7 service nova] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Refreshing network info cache for port 81591716-fb4a-427f-b3e9-a8b265fe1ce1 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 909.617987] env[62753]: DEBUG oslo_vmware.api [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332355, 'name': PowerOffVM_Task, 'duration_secs': 0.265112} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 909.618277] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 909.618510] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Volume detach. Driver type: vmdk {{(pid=62753) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 909.618741] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284612', 'volume_id': 'f9588f70-4cc6-40d4-9c38-0fda98e95283', 'name': 'volume-f9588f70-4cc6-40d4-9c38-0fda98e95283', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f03739f4-921f-46e5-b0e7-e9a99d96d621', 'attached_at': '', 'detached_at': '', 'volume_id': 'f9588f70-4cc6-40d4-9c38-0fda98e95283', 'serial': 'f9588f70-4cc6-40d4-9c38-0fda98e95283'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 909.619952] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-300222bf-51ee-4321-be3c-c9cc8950765c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.644332] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6a33800-7793-42ed-9e4b-7fad490dacc7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.651699] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f35a13a-1daf-45d0-808e-4d779efb32f7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.680798] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8508f4f7-ef01-4c1b-97ed-d90d7d1b6faa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.697064] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] The volume has not been displaced from its original location: [datastore1] volume-f9588f70-4cc6-40d4-9c38-0fda98e95283/volume-f9588f70-4cc6-40d4-9c38-0fda98e95283.vmdk. No consolidation needed. {{(pid=62753) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 909.702356] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Reconfiguring VM instance instance-00000045 to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 909.703235] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-12e872b2-7055-4906-b7fe-a16d6731ec36 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 909.724473] env[62753]: DEBUG oslo_vmware.api [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 909.724473] env[62753]: value = "task-1332356" [ 909.724473] env[62753]: _type = "Task" [ 909.724473] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 909.732202] env[62753]: DEBUG oslo_vmware.api [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332356, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 909.786488] env[62753]: DEBUG nova.compute.manager [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 909.894956] env[62753]: DEBUG nova.network.neutron [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 910.012542] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54b0aa8d-bc07-4364-a1c1-db3c294fde47 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.020484] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d44cb60-28d3-49d2-b583-de94cb7a8c69 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.052516] env[62753]: DEBUG nova.network.neutron [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Updating instance_info_cache with network_info: [{"id": "4037b488-2f26-40ee-b044-3b4fec79a454", "address": "fa:16:3e:6f:b1:75", "network": {"id": "56efb90b-64b4-484e-8062-eef8d39b1caf", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1294891206-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "909db6176c8f475caf2562a593bc61a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4037b488-2f", "ovs_interfaceid": "4037b488-2f26-40ee-b044-3b4fec79a454", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.054305] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be46932-79fe-418b-a6be-3caf292faaaf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.062700] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-855d5cf7-52b6-4582-8518-4984a5ba20d7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.077251] env[62753]: DEBUG nova.compute.provider_tree [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 910.185273] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "interface-3bc05fdb-81be-4764-b9d9-01acd125b020-9713beba-767d-408c-bd7d-bef58dc3c405" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.185273] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "interface-3bc05fdb-81be-4764-b9d9-01acd125b020-9713beba-767d-408c-bd7d-bef58dc3c405" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.247985] env[62753]: DEBUG oslo_vmware.api [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332356, 'name': ReconfigVM_Task, 'duration_secs': 0.188547} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.248398] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Reconfigured VM instance instance-00000045 to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 910.253280] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b0f46d7-016d-4246-b5df-cfbfc40fedc2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.271343] env[62753]: DEBUG oslo_vmware.api [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 910.271343] env[62753]: value = "task-1332357" [ 910.271343] env[62753]: _type = "Task" [ 910.271343] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.280755] env[62753]: DEBUG oslo_vmware.api [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332357, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.378851] env[62753]: DEBUG nova.network.neutron [req-1be711d0-4dc1-4674-b223-4ecca1af663a req-5d85b151-6c29-4fcf-92d2-ce5031ce1ad7 service nova] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updated VIF entry in instance network info cache for port 81591716-fb4a-427f-b3e9-a8b265fe1ce1. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 910.379237] env[62753]: DEBUG nova.network.neutron [req-1be711d0-4dc1-4674-b223-4ecca1af663a req-5d85b151-6c29-4fcf-92d2-ce5031ce1ad7 service nova] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating instance_info_cache with network_info: [{"id": "81591716-fb4a-427f-b3e9-a8b265fe1ce1", "address": "fa:16:3e:3e:ef:2c", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81591716-fb", "ovs_interfaceid": "81591716-fb4a-427f-b3e9-a8b265fe1ce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 910.559210] env[62753]: DEBUG oslo_concurrency.lockutils [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Releasing lock "refresh_cache-77cdd901-cc96-4a6f-a696-a65f54a96b1a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.559210] env[62753]: DEBUG nova.compute.manager [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Instance network_info: |[{"id": "4037b488-2f26-40ee-b044-3b4fec79a454", "address": "fa:16:3e:6f:b1:75", "network": {"id": "56efb90b-64b4-484e-8062-eef8d39b1caf", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1294891206-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "909db6176c8f475caf2562a593bc61a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4037b488-2f", "ovs_interfaceid": "4037b488-2f26-40ee-b044-3b4fec79a454", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 910.559660] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:b1:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8480e3f5-68bd-4c27-ae1f-7c994a8202b1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4037b488-2f26-40ee-b044-3b4fec79a454', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 910.567178] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Creating folder: Project (909db6176c8f475caf2562a593bc61a9). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 910.567480] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-da357969-f466-4544-8498-218b86495293 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.577839] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Created folder: Project (909db6176c8f475caf2562a593bc61a9) in parent group-v284541. [ 910.577989] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Creating folder: Instances. Parent ref: group-v284613. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 910.578264] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-53ef6674-e9c3-4802-a380-bc5d2725c297 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.580514] env[62753]: DEBUG nova.scheduler.client.report [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 910.591302] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Created folder: Instances in parent group-v284613. [ 910.591548] env[62753]: DEBUG oslo.service.loopingcall [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 910.591744] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 910.592287] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-574c1fc8-1d89-4555-9f60-1b745d194c06 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.612173] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 910.612173] env[62753]: value = "task-1332360" [ 910.612173] env[62753]: _type = "Task" [ 910.612173] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.620815] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332360, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.690215] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.690345] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.691293] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83029b80-c6b6-45a1-8260-801994f43372 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.709922] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-956c89ce-ff18-49aa-a2c0-c0b1d6746492 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.737816] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Reconfiguring VM to detach interface {{(pid=62753) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 910.738511] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f3f5a64b-0276-4db7-a5df-b0aebe98f1b6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.757374] env[62753]: DEBUG oslo_vmware.api [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 910.757374] env[62753]: value = "task-1332361" [ 910.757374] env[62753]: _type = "Task" [ 910.757374] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.767300] env[62753]: DEBUG oslo_vmware.api [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332361, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.780589] env[62753]: DEBUG oslo_vmware.api [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332357, 'name': ReconfigVM_Task, 'duration_secs': 0.174284} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 910.780900] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284612', 'volume_id': 'f9588f70-4cc6-40d4-9c38-0fda98e95283', 'name': 'volume-f9588f70-4cc6-40d4-9c38-0fda98e95283', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'f03739f4-921f-46e5-b0e7-e9a99d96d621', 'attached_at': '', 'detached_at': '', 'volume_id': 'f9588f70-4cc6-40d4-9c38-0fda98e95283', 'serial': 'f9588f70-4cc6-40d4-9c38-0fda98e95283'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 910.781228] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 910.781973] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f1e3752-db12-48d1-943e-ba59bb590cbb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.788780] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 910.789038] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7577d957-1a6a-44ac-bb4b-74f937a11ab6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.798994] env[62753]: DEBUG nova.compute.manager [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 910.830127] env[62753]: DEBUG nova.virt.hardware [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 910.830427] env[62753]: DEBUG nova.virt.hardware [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 910.830614] env[62753]: DEBUG nova.virt.hardware [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 910.830826] env[62753]: DEBUG nova.virt.hardware [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 910.831073] env[62753]: DEBUG nova.virt.hardware [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 910.831292] env[62753]: DEBUG nova.virt.hardware [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 910.831538] env[62753]: DEBUG nova.virt.hardware [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 910.831722] env[62753]: DEBUG nova.virt.hardware [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 910.831921] env[62753]: DEBUG nova.virt.hardware [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 910.832117] env[62753]: DEBUG nova.virt.hardware [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 910.832313] env[62753]: DEBUG nova.virt.hardware [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 910.833309] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c26a49c4-6db3-4054-9f0c-65f23e6030b1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.842138] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7507df6d-d616-448e-ba99-9ecee5a6d6a8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.860893] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 910.861163] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 910.861384] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Deleting the datastore file [datastore1] f03739f4-921f-46e5-b0e7-e9a99d96d621 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 910.861656] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-48098d91-ab7c-4468-9899-14f0b10932bb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 910.867277] env[62753]: DEBUG oslo_vmware.api [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 910.867277] env[62753]: value = "task-1332363" [ 910.867277] env[62753]: _type = "Task" [ 910.867277] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 910.874933] env[62753]: DEBUG oslo_vmware.api [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332363, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 910.882126] env[62753]: DEBUG oslo_concurrency.lockutils [req-1be711d0-4dc1-4674-b223-4ecca1af663a req-5d85b151-6c29-4fcf-92d2-ce5031ce1ad7 service nova] Releasing lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 910.988538] env[62753]: DEBUG nova.compute.manager [req-68bf9d3c-cb70-4565-8499-a491a1f17e2c req-2c65b35a-9566-4439-911d-1e7e65d4c0d2 service nova] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Received event network-vif-plugged-4037b488-2f26-40ee-b044-3b4fec79a454 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 910.988760] env[62753]: DEBUG oslo_concurrency.lockutils [req-68bf9d3c-cb70-4565-8499-a491a1f17e2c req-2c65b35a-9566-4439-911d-1e7e65d4c0d2 service nova] Acquiring lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 910.988978] env[62753]: DEBUG oslo_concurrency.lockutils [req-68bf9d3c-cb70-4565-8499-a491a1f17e2c req-2c65b35a-9566-4439-911d-1e7e65d4c0d2 service nova] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 910.989187] env[62753]: DEBUG oslo_concurrency.lockutils [req-68bf9d3c-cb70-4565-8499-a491a1f17e2c req-2c65b35a-9566-4439-911d-1e7e65d4c0d2 service nova] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 910.989453] env[62753]: DEBUG nova.compute.manager [req-68bf9d3c-cb70-4565-8499-a491a1f17e2c req-2c65b35a-9566-4439-911d-1e7e65d4c0d2 service nova] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] No waiting events found dispatching network-vif-plugged-4037b488-2f26-40ee-b044-3b4fec79a454 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 910.989696] env[62753]: WARNING nova.compute.manager [req-68bf9d3c-cb70-4565-8499-a491a1f17e2c req-2c65b35a-9566-4439-911d-1e7e65d4c0d2 service nova] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Received unexpected event network-vif-plugged-4037b488-2f26-40ee-b044-3b4fec79a454 for instance with vm_state building and task_state spawning. [ 910.989777] env[62753]: DEBUG nova.compute.manager [req-68bf9d3c-cb70-4565-8499-a491a1f17e2c req-2c65b35a-9566-4439-911d-1e7e65d4c0d2 service nova] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Received event network-changed-4037b488-2f26-40ee-b044-3b4fec79a454 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 910.989938] env[62753]: DEBUG nova.compute.manager [req-68bf9d3c-cb70-4565-8499-a491a1f17e2c req-2c65b35a-9566-4439-911d-1e7e65d4c0d2 service nova] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Refreshing instance network info cache due to event network-changed-4037b488-2f26-40ee-b044-3b4fec79a454. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 910.990191] env[62753]: DEBUG oslo_concurrency.lockutils [req-68bf9d3c-cb70-4565-8499-a491a1f17e2c req-2c65b35a-9566-4439-911d-1e7e65d4c0d2 service nova] Acquiring lock "refresh_cache-77cdd901-cc96-4a6f-a696-a65f54a96b1a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 910.990268] env[62753]: DEBUG oslo_concurrency.lockutils [req-68bf9d3c-cb70-4565-8499-a491a1f17e2c req-2c65b35a-9566-4439-911d-1e7e65d4c0d2 service nova] Acquired lock "refresh_cache-77cdd901-cc96-4a6f-a696-a65f54a96b1a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 910.990428] env[62753]: DEBUG nova.network.neutron [req-68bf9d3c-cb70-4565-8499-a491a1f17e2c req-2c65b35a-9566-4439-911d-1e7e65d4c0d2 service nova] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Refreshing network info cache for port 4037b488-2f26-40ee-b044-3b4fec79a454 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 911.049921] env[62753]: DEBUG nova.compute.manager [req-7de365e6-2fee-490a-ab7c-dee5cf536199 req-6a213d9f-0084-4121-9047-4eed3035799a service nova] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Received event network-vif-plugged-ef993aa6-10f5-4de9-b652-426dbaac5112 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 911.050192] env[62753]: DEBUG oslo_concurrency.lockutils [req-7de365e6-2fee-490a-ab7c-dee5cf536199 req-6a213d9f-0084-4121-9047-4eed3035799a service nova] Acquiring lock "d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 911.050416] env[62753]: DEBUG oslo_concurrency.lockutils [req-7de365e6-2fee-490a-ab7c-dee5cf536199 req-6a213d9f-0084-4121-9047-4eed3035799a service nova] Lock "d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.050589] env[62753]: DEBUG oslo_concurrency.lockutils [req-7de365e6-2fee-490a-ab7c-dee5cf536199 req-6a213d9f-0084-4121-9047-4eed3035799a service nova] Lock "d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.050761] env[62753]: DEBUG nova.compute.manager [req-7de365e6-2fee-490a-ab7c-dee5cf536199 req-6a213d9f-0084-4121-9047-4eed3035799a service nova] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] No waiting events found dispatching network-vif-plugged-ef993aa6-10f5-4de9-b652-426dbaac5112 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 911.050931] env[62753]: WARNING nova.compute.manager [req-7de365e6-2fee-490a-ab7c-dee5cf536199 req-6a213d9f-0084-4121-9047-4eed3035799a service nova] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Received unexpected event network-vif-plugged-ef993aa6-10f5-4de9-b652-426dbaac5112 for instance with vm_state building and task_state spawning. [ 911.086408] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.311s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 911.086911] env[62753]: DEBUG nova.compute.manager [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 911.089369] env[62753]: DEBUG oslo_concurrency.lockutils [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 20.497s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 911.090757] env[62753]: INFO nova.compute.claims [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 911.122983] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332360, 'name': CreateVM_Task, 'duration_secs': 0.374716} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.122983] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 911.123094] env[62753]: DEBUG oslo_concurrency.lockutils [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.123269] env[62753]: DEBUG oslo_concurrency.lockutils [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.123620] env[62753]: DEBUG oslo_concurrency.lockutils [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 911.123860] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cd1a7d24-1bf6-4da6-9efb-574bbede695b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.129818] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 911.129818] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5227df39-83b9-4b6a-31db-881af7e0957d" [ 911.129818] env[62753]: _type = "Task" [ 911.129818] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.138173] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5227df39-83b9-4b6a-31db-881af7e0957d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.267876] env[62753]: DEBUG oslo_vmware.api [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332361, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.376857] env[62753]: DEBUG oslo_vmware.api [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332363, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.186724} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.377126] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 911.377320] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 911.377500] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 911.377675] env[62753]: INFO nova.compute.manager [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Took 2.28 seconds to destroy the instance on the hypervisor. [ 911.377912] env[62753]: DEBUG oslo.service.loopingcall [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 911.378117] env[62753]: DEBUG nova.compute.manager [-] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 911.378215] env[62753]: DEBUG nova.network.neutron [-] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 911.596283] env[62753]: DEBUG nova.compute.utils [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 911.607152] env[62753]: DEBUG nova.compute.manager [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 911.607152] env[62753]: DEBUG nova.network.neutron [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 911.637936] env[62753]: DEBUG nova.network.neutron [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Successfully updated port: ef993aa6-10f5-4de9-b652-426dbaac5112 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 911.647938] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5227df39-83b9-4b6a-31db-881af7e0957d, 'name': SearchDatastore_Task, 'duration_secs': 0.011683} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 911.648310] env[62753]: DEBUG oslo_concurrency.lockutils [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 911.648552] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 911.648787] env[62753]: DEBUG oslo_concurrency.lockutils [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 911.649424] env[62753]: DEBUG oslo_concurrency.lockutils [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 911.649547] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 911.653757] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e34db419-6edc-4d6d-8f71-acd85640020f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.672213] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 911.672417] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 911.673493] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6f849fdb-8e29-4a43-a600-4793db5cf3c2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 911.679831] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 911.679831] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52642a41-90f0-e486-8ca3-86407d85aafb" [ 911.679831] env[62753]: _type = "Task" [ 911.679831] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 911.690773] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52642a41-90f0-e486-8ca3-86407d85aafb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.725956] env[62753]: DEBUG nova.policy [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cbc9dc744e144ff28cf4c44fb043094f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e376836411ae4854965adf2923eaaffd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 911.768456] env[62753]: DEBUG oslo_vmware.api [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332361, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 911.961440] env[62753]: DEBUG nova.network.neutron [req-68bf9d3c-cb70-4565-8499-a491a1f17e2c req-2c65b35a-9566-4439-911d-1e7e65d4c0d2 service nova] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Updated VIF entry in instance network info cache for port 4037b488-2f26-40ee-b044-3b4fec79a454. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 911.961736] env[62753]: DEBUG nova.network.neutron [req-68bf9d3c-cb70-4565-8499-a491a1f17e2c req-2c65b35a-9566-4439-911d-1e7e65d4c0d2 service nova] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Updating instance_info_cache with network_info: [{"id": "4037b488-2f26-40ee-b044-3b4fec79a454", "address": "fa:16:3e:6f:b1:75", "network": {"id": "56efb90b-64b4-484e-8062-eef8d39b1caf", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1294891206-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "909db6176c8f475caf2562a593bc61a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4037b488-2f", "ovs_interfaceid": "4037b488-2f26-40ee-b044-3b4fec79a454", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.107897] env[62753]: DEBUG nova.compute.manager [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 912.134715] env[62753]: DEBUG nova.network.neutron [-] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.140431] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Acquiring lock "refresh_cache-d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 912.141559] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Acquired lock "refresh_cache-d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 912.141559] env[62753]: DEBUG nova.network.neutron [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 912.192488] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52642a41-90f0-e486-8ca3-86407d85aafb, 'name': SearchDatastore_Task, 'duration_secs': 0.037725} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.195634] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ea02b52b-da04-46e4-8dcb-396e95934440 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.200995] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 912.200995] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a3f46d-dce0-6b66-81d3-c6a392b5ce0f" [ 912.200995] env[62753]: _type = "Task" [ 912.200995] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.208968] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a3f46d-dce0-6b66-81d3-c6a392b5ce0f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.263127] env[62753]: DEBUG nova.network.neutron [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Successfully created port: d17809a7-cdfb-44c8-9b83-1d5110ede0d3 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 912.271926] env[62753]: DEBUG oslo_vmware.api [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332361, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.359214] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95aded18-8516-4812-8a28-b2f024e1546e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.367097] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4849620d-c8fb-436c-885e-36dee5902df4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.400406] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ebf1c933-49e0-4a69-b38d-beb8891b060f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.408152] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-752a8700-cc25-4eb8-845a-d59c1bbe05de {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.412088] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 912.412369] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 912.412587] env[62753]: INFO nova.compute.manager [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Shelving [ 912.425937] env[62753]: DEBUG nova.compute.provider_tree [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 912.464515] env[62753]: DEBUG oslo_concurrency.lockutils [req-68bf9d3c-cb70-4565-8499-a491a1f17e2c req-2c65b35a-9566-4439-911d-1e7e65d4c0d2 service nova] Releasing lock "refresh_cache-77cdd901-cc96-4a6f-a696-a65f54a96b1a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.640565] env[62753]: INFO nova.compute.manager [-] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Took 1.26 seconds to deallocate network for instance. [ 912.677915] env[62753]: DEBUG nova.network.neutron [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 912.711741] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a3f46d-dce0-6b66-81d3-c6a392b5ce0f, 'name': SearchDatastore_Task, 'duration_secs': 0.036656} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 912.713051] env[62753]: DEBUG oslo_concurrency.lockutils [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 912.713051] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 77cdd901-cc96-4a6f-a696-a65f54a96b1a/77cdd901-cc96-4a6f-a696-a65f54a96b1a.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 912.713051] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-5568868e-2fd9-4007-b17b-29850853c0ff {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.719036] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 912.719036] env[62753]: value = "task-1332364" [ 912.719036] env[62753]: _type = "Task" [ 912.719036] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.726649] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332364, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.768996] env[62753]: DEBUG oslo_vmware.api [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332361, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 912.851312] env[62753]: DEBUG nova.network.neutron [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Updating instance_info_cache with network_info: [{"id": "ef993aa6-10f5-4de9-b652-426dbaac5112", "address": "fa:16:3e:37:0d:b5", "network": {"id": "978c6598-cb9f-4d90-9ccb-1de3097a91a6", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-110846568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2dc6fe36ae84435a8e09cdf56340c58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1323cb03-8367-485a-962e-131af8eba474", "external-id": "nsx-vlan-transportzone-41", "segmentation_id": 41, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef993aa6-10", "ovs_interfaceid": "ef993aa6-10f5-4de9-b652-426dbaac5112", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 912.920917] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 912.921685] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9ee8e80b-33ed-42ad-aa16-822e9e83d066 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 912.928668] env[62753]: DEBUG nova.scheduler.client.report [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 912.934804] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 912.934804] env[62753]: value = "task-1332365" [ 912.934804] env[62753]: _type = "Task" [ 912.934804] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 912.946362] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332365, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.024987] env[62753]: DEBUG nova.compute.manager [req-9d61b4e7-23a8-4282-9ec7-419958a12733 req-5611911d-bcc9-4de2-86b8-68affb0e5a43 service nova] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Received event network-vif-deleted-ce484d8b-65ec-4512-aaf1-48a0c51f4f74 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.025191] env[62753]: DEBUG nova.compute.manager [req-9d61b4e7-23a8-4282-9ec7-419958a12733 req-5611911d-bcc9-4de2-86b8-68affb0e5a43 service nova] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Received event network-changed-ef993aa6-10f5-4de9-b652-426dbaac5112 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.025367] env[62753]: DEBUG nova.compute.manager [req-9d61b4e7-23a8-4282-9ec7-419958a12733 req-5611911d-bcc9-4de2-86b8-68affb0e5a43 service nova] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Refreshing instance network info cache due to event network-changed-ef993aa6-10f5-4de9-b652-426dbaac5112. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 913.025624] env[62753]: DEBUG oslo_concurrency.lockutils [req-9d61b4e7-23a8-4282-9ec7-419958a12733 req-5611911d-bcc9-4de2-86b8-68affb0e5a43 service nova] Acquiring lock "refresh_cache-d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.114985] env[62753]: DEBUG nova.compute.manager [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 913.143225] env[62753]: DEBUG nova.virt.hardware [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 913.143543] env[62753]: DEBUG nova.virt.hardware [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 913.143724] env[62753]: DEBUG nova.virt.hardware [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 913.143906] env[62753]: DEBUG nova.virt.hardware [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 913.144154] env[62753]: DEBUG nova.virt.hardware [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 913.144349] env[62753]: DEBUG nova.virt.hardware [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 913.144571] env[62753]: DEBUG nova.virt.hardware [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 913.144746] env[62753]: DEBUG nova.virt.hardware [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 913.144968] env[62753]: DEBUG nova.virt.hardware [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 913.145170] env[62753]: DEBUG nova.virt.hardware [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 913.145360] env[62753]: DEBUG nova.virt.hardware [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 913.146280] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-882344a2-49b6-4cf8-83a7-ba133d756ecd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.155615] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed3a62a0-2912-47c7-9304-76d7351978be {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.201535] env[62753]: INFO nova.compute.manager [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Took 0.56 seconds to detach 1 volumes for instance. [ 913.229037] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332364, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.270730] env[62753]: DEBUG oslo_vmware.api [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332361, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.354930] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Releasing lock "refresh_cache-d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 913.355828] env[62753]: DEBUG nova.compute.manager [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Instance network_info: |[{"id": "ef993aa6-10f5-4de9-b652-426dbaac5112", "address": "fa:16:3e:37:0d:b5", "network": {"id": "978c6598-cb9f-4d90-9ccb-1de3097a91a6", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-110846568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2dc6fe36ae84435a8e09cdf56340c58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1323cb03-8367-485a-962e-131af8eba474", "external-id": "nsx-vlan-transportzone-41", "segmentation_id": 41, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef993aa6-10", "ovs_interfaceid": "ef993aa6-10f5-4de9-b652-426dbaac5112", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 913.355988] env[62753]: DEBUG oslo_concurrency.lockutils [req-9d61b4e7-23a8-4282-9ec7-419958a12733 req-5611911d-bcc9-4de2-86b8-68affb0e5a43 service nova] Acquired lock "refresh_cache-d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.356368] env[62753]: DEBUG nova.network.neutron [req-9d61b4e7-23a8-4282-9ec7-419958a12733 req-5611911d-bcc9-4de2-86b8-68affb0e5a43 service nova] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Refreshing network info cache for port ef993aa6-10f5-4de9-b652-426dbaac5112 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 913.358311] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:37:0d:b5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '1323cb03-8367-485a-962e-131af8eba474', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ef993aa6-10f5-4de9-b652-426dbaac5112', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 913.368993] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Creating folder: Project (c2dc6fe36ae84435a8e09cdf56340c58). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 913.370115] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-2b2ab46b-0863-4733-b52d-4769296d6a18 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.380901] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Created folder: Project (c2dc6fe36ae84435a8e09cdf56340c58) in parent group-v284541. [ 913.381140] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Creating folder: Instances. Parent ref: group-v284616. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 913.381400] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-be098ebb-1249-4f07-b978-3416e2c2432c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.392208] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Created folder: Instances in parent group-v284616. [ 913.392525] env[62753]: DEBUG oslo.service.loopingcall [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 913.392723] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 913.393208] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-1caa20e3-d448-4b73-8554-9e1e93355767 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.412050] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 913.412050] env[62753]: value = "task-1332368" [ 913.412050] env[62753]: _type = "Task" [ 913.412050] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.422902] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332368, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.436183] env[62753]: DEBUG oslo_concurrency.lockutils [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.347s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.436729] env[62753]: DEBUG nova.compute.manager [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 913.440224] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 21.350s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.441255] env[62753]: INFO nova.compute.claims [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 913.452833] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332365, 'name': PowerOffVM_Task, 'duration_secs': 0.193926} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.453221] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 913.454106] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21fa3c5f-d49f-4246-8125-31d0c358b4af {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.473439] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-899bb02e-f504-40f5-b9f7-cd252894dcfb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.684555] env[62753]: DEBUG nova.compute.manager [req-0b082525-2b1b-4c11-bc5f-487a0933aa92 req-45ea4f40-6eec-4501-bb9d-86c020ca998d service nova] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Received event network-vif-plugged-d17809a7-cdfb-44c8-9b83-1d5110ede0d3 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 913.684555] env[62753]: DEBUG oslo_concurrency.lockutils [req-0b082525-2b1b-4c11-bc5f-487a0933aa92 req-45ea4f40-6eec-4501-bb9d-86c020ca998d service nova] Acquiring lock "b5674964-1928-4ecb-b1fd-8f60a94b4270-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.685182] env[62753]: DEBUG oslo_concurrency.lockutils [req-0b082525-2b1b-4c11-bc5f-487a0933aa92 req-45ea4f40-6eec-4501-bb9d-86c020ca998d service nova] Lock "b5674964-1928-4ecb-b1fd-8f60a94b4270-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 913.685182] env[62753]: DEBUG oslo_concurrency.lockutils [req-0b082525-2b1b-4c11-bc5f-487a0933aa92 req-45ea4f40-6eec-4501-bb9d-86c020ca998d service nova] Lock "b5674964-1928-4ecb-b1fd-8f60a94b4270-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 913.685182] env[62753]: DEBUG nova.compute.manager [req-0b082525-2b1b-4c11-bc5f-487a0933aa92 req-45ea4f40-6eec-4501-bb9d-86c020ca998d service nova] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] No waiting events found dispatching network-vif-plugged-d17809a7-cdfb-44c8-9b83-1d5110ede0d3 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 913.685301] env[62753]: WARNING nova.compute.manager [req-0b082525-2b1b-4c11-bc5f-487a0933aa92 req-45ea4f40-6eec-4501-bb9d-86c020ca998d service nova] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Received unexpected event network-vif-plugged-d17809a7-cdfb-44c8-9b83-1d5110ede0d3 for instance with vm_state building and task_state spawning. [ 913.707887] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 913.729241] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332364, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.511237} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.729479] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 77cdd901-cc96-4a6f-a696-a65f54a96b1a/77cdd901-cc96-4a6f-a696-a65f54a96b1a.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 913.729688] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 913.729922] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-3b1450e7-cfa4-4036-8548-1c6b7618ec20 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.735788] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 913.735788] env[62753]: value = "task-1332369" [ 913.735788] env[62753]: _type = "Task" [ 913.735788] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.743202] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332369, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.770157] env[62753]: DEBUG oslo_vmware.api [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332361, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.771628] env[62753]: DEBUG nova.network.neutron [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Successfully updated port: d17809a7-cdfb-44c8-9b83-1d5110ede0d3 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 913.921824] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332368, 'name': CreateVM_Task, 'duration_secs': 0.317417} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 913.921984] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 913.922653] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 913.922824] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 913.923181] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 913.923434] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-413231fa-7a0f-4bbc-91da-20de29cc111e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.927556] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Waiting for the task: (returnval){ [ 913.927556] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5235d279-6349-e609-9812-a43938bc23cb" [ 913.927556] env[62753]: _type = "Task" [ 913.927556] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 913.935879] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5235d279-6349-e609-9812-a43938bc23cb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 913.949101] env[62753]: DEBUG nova.compute.utils [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 913.950254] env[62753]: DEBUG nova.compute.manager [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 913.950404] env[62753]: DEBUG nova.network.neutron [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 913.984116] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Creating Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 913.984430] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-4cb58ecd-3f8d-4dac-bce0-a449b5fafa29 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 913.991774] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 913.991774] env[62753]: value = "task-1332370" [ 913.991774] env[62753]: _type = "Task" [ 913.991774] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.000238] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332370, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.034916] env[62753]: DEBUG nova.policy [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b6207815e8041b4bba3069635d7d962', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6c8a6bc946a469fa85cc3ab80d7333a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 914.246394] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332369, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064833} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.246668] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 914.247664] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7513a805-52a7-49aa-bbdc-18f7d507a42a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.270532] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] 77cdd901-cc96-4a6f-a696-a65f54a96b1a/77cdd901-cc96-4a6f-a696-a65f54a96b1a.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 914.274788] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1656d224-7082-4f01-8563-1a39c536a75f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.299278] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "refresh_cache-b5674964-1928-4ecb-b1fd-8f60a94b4270" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.299423] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired lock "refresh_cache-b5674964-1928-4ecb-b1fd-8f60a94b4270" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.299790] env[62753]: DEBUG nova.network.neutron [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 914.308025] env[62753]: DEBUG oslo_vmware.api [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332361, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.309925] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 914.309925] env[62753]: value = "task-1332371" [ 914.309925] env[62753]: _type = "Task" [ 914.309925] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.318260] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332371, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.320143] env[62753]: DEBUG nova.network.neutron [req-9d61b4e7-23a8-4282-9ec7-419958a12733 req-5611911d-bcc9-4de2-86b8-68affb0e5a43 service nova] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Updated VIF entry in instance network info cache for port ef993aa6-10f5-4de9-b652-426dbaac5112. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 914.320475] env[62753]: DEBUG nova.network.neutron [req-9d61b4e7-23a8-4282-9ec7-419958a12733 req-5611911d-bcc9-4de2-86b8-68affb0e5a43 service nova] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Updating instance_info_cache with network_info: [{"id": "ef993aa6-10f5-4de9-b652-426dbaac5112", "address": "fa:16:3e:37:0d:b5", "network": {"id": "978c6598-cb9f-4d90-9ccb-1de3097a91a6", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-110846568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2dc6fe36ae84435a8e09cdf56340c58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1323cb03-8367-485a-962e-131af8eba474", "external-id": "nsx-vlan-transportzone-41", "segmentation_id": 41, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef993aa6-10", "ovs_interfaceid": "ef993aa6-10f5-4de9-b652-426dbaac5112", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 914.439086] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5235d279-6349-e609-9812-a43938bc23cb, 'name': SearchDatastore_Task, 'duration_secs': 0.053321} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.439564] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.439840] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 914.440109] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 914.440281] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 914.440478] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 914.440759] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-cb5ec74b-3ad1-4f48-ac48-184e9b9130f5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.451854] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 914.452083] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 914.453326] env[62753]: DEBUG nova.compute.manager [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 914.459927] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7abe34cd-b2bb-4e9f-8444-9b4803a5aa5a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.470800] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Waiting for the task: (returnval){ [ 914.470800] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e0e13f-ddf1-4050-eb3a-807ab8f875e6" [ 914.470800] env[62753]: _type = "Task" [ 914.470800] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.482305] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e0e13f-ddf1-4050-eb3a-807ab8f875e6, 'name': SearchDatastore_Task, 'duration_secs': 0.009881} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.483832] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e8a04207-4cc5-4117-a2ff-a60abb6f087b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.491228] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Waiting for the task: (returnval){ [ 914.491228] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52faecd7-1387-d41c-3e4f-af44f1cb77a9" [ 914.491228] env[62753]: _type = "Task" [ 914.491228] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.503465] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52faecd7-1387-d41c-3e4f-af44f1cb77a9, 'name': SearchDatastore_Task, 'duration_secs': 0.008986} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.508753] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.509097] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075/d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 914.509459] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332370, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.509806] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2b7cb7b2-4695-4cf7-a366-818370ec0f36 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.516625] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Waiting for the task: (returnval){ [ 914.516625] env[62753]: value = "task-1332372" [ 914.516625] env[62753]: _type = "Task" [ 914.516625] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.527406] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': task-1332372, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.673392] env[62753]: DEBUG nova.network.neutron [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Successfully created port: 7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 914.748895] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c80fe1a1-3df3-45e3-bd72-e68951f62fed {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.757357] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df7c5378-fe79-4146-91c4-7b48e4d25293 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.794899] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e8ab0c1-7066-43c0-9486-979c11465692 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.797850] env[62753]: DEBUG oslo_vmware.api [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332361, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.805755] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-988e5eec-2c2f-4e8f-86ad-c706a03fda61 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.819899] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332371, 'name': ReconfigVM_Task, 'duration_secs': 0.28388} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 914.828023] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Reconfigured VM instance instance-00000049 to attach disk [datastore2] 77cdd901-cc96-4a6f-a696-a65f54a96b1a/77cdd901-cc96-4a6f-a696-a65f54a96b1a.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 914.828991] env[62753]: DEBUG oslo_concurrency.lockutils [req-9d61b4e7-23a8-4282-9ec7-419958a12733 req-5611911d-bcc9-4de2-86b8-68affb0e5a43 service nova] Releasing lock "refresh_cache-d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 914.829673] env[62753]: DEBUG nova.compute.provider_tree [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 914.830937] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d1a840fa-f0d2-44d8-9f5d-4eeea60c1872 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 914.837960] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 914.837960] env[62753]: value = "task-1332373" [ 914.837960] env[62753]: _type = "Task" [ 914.837960] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 914.848717] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332373, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 914.858775] env[62753]: DEBUG nova.network.neutron [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 915.006069] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332370, 'name': CreateSnapshot_Task, 'duration_secs': 0.624045} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.006985] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Created Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 915.007109] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50f86f04-51f5-4735-adfc-54e1f729ce1e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.027550] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': task-1332372, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.480924} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.027684] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075/d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 915.027903] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 915.028180] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-37bbe1f7-ea37-4d41-8ff3-2bfb31697c3d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.035597] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Waiting for the task: (returnval){ [ 915.035597] env[62753]: value = "task-1332374" [ 915.035597] env[62753]: _type = "Task" [ 915.035597] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.045931] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': task-1332374, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.113546] env[62753]: DEBUG nova.network.neutron [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Updating instance_info_cache with network_info: [{"id": "d17809a7-cdfb-44c8-9b83-1d5110ede0d3", "address": "fa:16:3e:fc:b3:fc", "network": {"id": "8c5edc3b-7faf-47f2-9736-7093f9ff72de", "bridge": "br-int", "label": "tempest-ImagesTestJSON-588052284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e376836411ae4854965adf2923eaaffd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76f377cd-5966-49b4-9210-907f592c694e", "external-id": "nsx-vlan-transportzone-124", "segmentation_id": 124, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd17809a7-cd", "ovs_interfaceid": "d17809a7-cdfb-44c8-9b83-1d5110ede0d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 915.273021] env[62753]: DEBUG oslo_vmware.api [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332361, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.334458] env[62753]: DEBUG nova.scheduler.client.report [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 915.351169] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332373, 'name': Rename_Task, 'duration_secs': 0.20759} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.352276] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 915.352607] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-06c36fed-149e-4721-888d-38a6ecbe4372 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.362251] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 915.362251] env[62753]: value = "task-1332375" [ 915.362251] env[62753]: _type = "Task" [ 915.362251] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.370919] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332375, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.474199] env[62753]: DEBUG nova.compute.manager [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 915.504489] env[62753]: DEBUG nova.virt.hardware [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 915.504764] env[62753]: DEBUG nova.virt.hardware [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 915.504944] env[62753]: DEBUG nova.virt.hardware [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 915.505172] env[62753]: DEBUG nova.virt.hardware [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 915.505330] env[62753]: DEBUG nova.virt.hardware [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 915.505496] env[62753]: DEBUG nova.virt.hardware [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 915.505727] env[62753]: DEBUG nova.virt.hardware [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 915.505897] env[62753]: DEBUG nova.virt.hardware [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 915.506086] env[62753]: DEBUG nova.virt.hardware [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 915.506260] env[62753]: DEBUG nova.virt.hardware [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 915.506441] env[62753]: DEBUG nova.virt.hardware [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 915.507407] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4949fbc8-66f9-423e-b502-90751bc15ef2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.515601] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21ad0476-f0bd-49e7-bd87-2ce317b0f327 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.535102] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Creating linked-clone VM from snapshot {{(pid=62753) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 915.536047] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f23b23e4-6781-4521-93be-c8b4460f3d39 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.547702] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': task-1332374, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060928} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 915.549074] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 915.549431] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 915.549431] env[62753]: value = "task-1332376" [ 915.549431] env[62753]: _type = "Task" [ 915.549431] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.550122] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-366e4436-b6b5-44fd-9ae3-20d29733310e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.560509] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332376, 'name': CloneVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.578191] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Reconfiguring VM instance instance-0000004a to attach disk [datastore2] d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075/d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 915.578518] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2ba10948-87cc-4c13-97e5-8ece6df3f5a1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.598860] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Waiting for the task: (returnval){ [ 915.598860] env[62753]: value = "task-1332377" [ 915.598860] env[62753]: _type = "Task" [ 915.598860] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.606669] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': task-1332377, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.616558] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Releasing lock "refresh_cache-b5674964-1928-4ecb-b1fd-8f60a94b4270" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 915.616970] env[62753]: DEBUG nova.compute.manager [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Instance network_info: |[{"id": "d17809a7-cdfb-44c8-9b83-1d5110ede0d3", "address": "fa:16:3e:fc:b3:fc", "network": {"id": "8c5edc3b-7faf-47f2-9736-7093f9ff72de", "bridge": "br-int", "label": "tempest-ImagesTestJSON-588052284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e376836411ae4854965adf2923eaaffd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76f377cd-5966-49b4-9210-907f592c694e", "external-id": "nsx-vlan-transportzone-124", "segmentation_id": 124, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd17809a7-cd", "ovs_interfaceid": "d17809a7-cdfb-44c8-9b83-1d5110ede0d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 915.617596] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:fc:b3:fc', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '76f377cd-5966-49b4-9210-907f592c694e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd17809a7-cdfb-44c8-9b83-1d5110ede0d3', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 915.626888] env[62753]: DEBUG oslo.service.loopingcall [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 915.627690] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 915.628012] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-a61d8833-74ac-4556-aff8-0e6209e6c695 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 915.649195] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 915.649195] env[62753]: value = "task-1332378" [ 915.649195] env[62753]: _type = "Task" [ 915.649195] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 915.658446] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332378, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.768871] env[62753]: DEBUG nova.compute.manager [req-742cd39a-16fe-4d9b-ad04-7757fdba7122 req-e121c24e-baf2-4115-b276-355e439e3934 service nova] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Received event network-changed-d17809a7-cdfb-44c8-9b83-1d5110ede0d3 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 915.769383] env[62753]: DEBUG nova.compute.manager [req-742cd39a-16fe-4d9b-ad04-7757fdba7122 req-e121c24e-baf2-4115-b276-355e439e3934 service nova] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Refreshing instance network info cache due to event network-changed-d17809a7-cdfb-44c8-9b83-1d5110ede0d3. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 915.769705] env[62753]: DEBUG oslo_concurrency.lockutils [req-742cd39a-16fe-4d9b-ad04-7757fdba7122 req-e121c24e-baf2-4115-b276-355e439e3934 service nova] Acquiring lock "refresh_cache-b5674964-1928-4ecb-b1fd-8f60a94b4270" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 915.769964] env[62753]: DEBUG oslo_concurrency.lockutils [req-742cd39a-16fe-4d9b-ad04-7757fdba7122 req-e121c24e-baf2-4115-b276-355e439e3934 service nova] Acquired lock "refresh_cache-b5674964-1928-4ecb-b1fd-8f60a94b4270" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 915.770265] env[62753]: DEBUG nova.network.neutron [req-742cd39a-16fe-4d9b-ad04-7757fdba7122 req-e121c24e-baf2-4115-b276-355e439e3934 service nova] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Refreshing network info cache for port d17809a7-cdfb-44c8-9b83-1d5110ede0d3 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 915.777433] env[62753]: DEBUG oslo_vmware.api [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332361, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 915.840840] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.401s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 915.841641] env[62753]: DEBUG nova.compute.manager [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 915.844604] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 20.579s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 915.844825] env[62753]: DEBUG nova.objects.instance [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lazy-loading 'resources' on Instance uuid 49311ef5-733a-4146-87ac-876a121a8d4d {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 915.873926] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332375, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.064107] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332376, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.109365] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': task-1332377, 'name': ReconfigVM_Task, 'duration_secs': 0.28304} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.109793] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Reconfigured VM instance instance-0000004a to attach disk [datastore2] d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075/d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 916.110570] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-858996d1-6f75-4b42-837e-c0653d1f8bc5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.118324] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Waiting for the task: (returnval){ [ 916.118324] env[62753]: value = "task-1332379" [ 916.118324] env[62753]: _type = "Task" [ 916.118324] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.129044] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': task-1332379, 'name': Rename_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.159253] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332378, 'name': CreateVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.278230] env[62753]: DEBUG oslo_vmware.api [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332361, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.350689] env[62753]: DEBUG nova.compute.utils [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 916.352351] env[62753]: DEBUG nova.compute.manager [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 916.352529] env[62753]: DEBUG nova.network.neutron [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 916.376895] env[62753]: DEBUG oslo_vmware.api [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332375, 'name': PowerOnVM_Task, 'duration_secs': 0.984831} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.376895] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 916.376895] env[62753]: INFO nova.compute.manager [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Took 7.90 seconds to spawn the instance on the hypervisor. [ 916.377043] env[62753]: DEBUG nova.compute.manager [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 916.377785] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-161e511e-ca6e-4336-a819-6a5cefbd0e5a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.412370] env[62753]: DEBUG nova.policy [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'a6ec6b26d1d5444e93c410ee4e2ea806', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '371a5e8b16be4cffbdeb2139757684fc', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 916.568511] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332376, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.600483] env[62753]: DEBUG nova.network.neutron [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Successfully updated port: 7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 916.625889] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f17db0-6cde-4fa0-8cb0-b5be43284e1c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.635652] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': task-1332379, 'name': Rename_Task, 'duration_secs': 0.255081} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.637521] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 916.637818] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3f9eca79-236b-4be1-8640-cfd4d14bb023 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.640176] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca89e6ff-769a-4b37-ae39-d59d4213979b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.682026] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-58a2421e-b537-4eba-8740-81cdd20e5cd2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.684988] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Waiting for the task: (returnval){ [ 916.684988] env[62753]: value = "task-1332380" [ 916.684988] env[62753]: _type = "Task" [ 916.684988] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.694236] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332378, 'name': CreateVM_Task, 'duration_secs': 0.712629} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.694576] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 916.696063] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-701b1aa8-0c2f-4a7e-8f02-661455aea4bc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.704108] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 916.704108] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 916.705293] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 916.705293] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': task-1332380, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.705293] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-85efbdb2-93d8-49df-b838-e641dd87bde0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 916.707707] env[62753]: DEBUG nova.network.neutron [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Successfully created port: 91add90d-924f-43de-bead-d4f33e2af4ad {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 916.718573] env[62753]: DEBUG nova.compute.provider_tree [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 916.720891] env[62753]: DEBUG nova.network.neutron [req-742cd39a-16fe-4d9b-ad04-7757fdba7122 req-e121c24e-baf2-4115-b276-355e439e3934 service nova] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Updated VIF entry in instance network info cache for port d17809a7-cdfb-44c8-9b83-1d5110ede0d3. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 916.721779] env[62753]: DEBUG nova.network.neutron [req-742cd39a-16fe-4d9b-ad04-7757fdba7122 req-e121c24e-baf2-4115-b276-355e439e3934 service nova] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Updating instance_info_cache with network_info: [{"id": "d17809a7-cdfb-44c8-9b83-1d5110ede0d3", "address": "fa:16:3e:fc:b3:fc", "network": {"id": "8c5edc3b-7faf-47f2-9736-7093f9ff72de", "bridge": "br-int", "label": "tempest-ImagesTestJSON-588052284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e376836411ae4854965adf2923eaaffd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76f377cd-5966-49b4-9210-907f592c694e", "external-id": "nsx-vlan-transportzone-124", "segmentation_id": 124, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd17809a7-cd", "ovs_interfaceid": "d17809a7-cdfb-44c8-9b83-1d5110ede0d3", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 916.728564] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 916.728564] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529ddf6c-2ec7-cc8b-7dd8-eb9016a679d4" [ 916.728564] env[62753]: _type = "Task" [ 916.728564] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 916.729988] env[62753]: DEBUG nova.scheduler.client.report [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 916.744659] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529ddf6c-2ec7-cc8b-7dd8-eb9016a679d4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 916.773652] env[62753]: DEBUG oslo_vmware.api [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332361, 'name': ReconfigVM_Task, 'duration_secs': 5.757716} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 916.773901] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 916.774168] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Reconfigured VM to detach interface {{(pid=62753) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 916.855694] env[62753]: DEBUG nova.compute.manager [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 916.906125] env[62753]: INFO nova.compute.manager [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Took 27.97 seconds to build instance. [ 917.067073] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332376, 'name': CloneVM_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.106990] env[62753]: DEBUG oslo_concurrency.lockutils [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "refresh_cache-a401909c-5978-492b-a4fa-b17d38a5af44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.107233] env[62753]: DEBUG oslo_concurrency.lockutils [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "refresh_cache-a401909c-5978-492b-a4fa-b17d38a5af44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.107450] env[62753]: DEBUG nova.network.neutron [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 917.195305] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': task-1332380, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.234650] env[62753]: DEBUG oslo_concurrency.lockutils [req-742cd39a-16fe-4d9b-ad04-7757fdba7122 req-e121c24e-baf2-4115-b276-355e439e3934 service nova] Releasing lock "refresh_cache-b5674964-1928-4ecb-b1fd-8f60a94b4270" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.235514] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.391s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.241048] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 14.834s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.250489] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529ddf6c-2ec7-cc8b-7dd8-eb9016a679d4, 'name': SearchDatastore_Task, 'duration_secs': 0.028794} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.250792] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 917.251039] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 917.251795] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.251795] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 917.251795] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 917.251950] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d0c9b81a-d786-45f3-9886-868f429fb356 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.261968] env[62753]: INFO nova.scheduler.client.report [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Deleted allocations for instance 49311ef5-733a-4146-87ac-876a121a8d4d [ 917.266368] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 917.266555] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 917.267878] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c958940c-8c73-4fd2-b9ef-011282249346 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.274233] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 917.274233] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]524e2667-ac62-92ff-b9c4-12e0fdbac39f" [ 917.274233] env[62753]: _type = "Task" [ 917.274233] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.282222] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]524e2667-ac62-92ff-b9c4-12e0fdbac39f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.408849] env[62753]: DEBUG oslo_concurrency.lockutils [None req-868b6ff4-36f1-41c2-9a93-f66d20d929d3 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 54.404s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.567568] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332376, 'name': CloneVM_Task, 'duration_secs': 1.601429} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.567850] env[62753]: INFO nova.virt.vmwareapi.vmops [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Created linked-clone VM from snapshot [ 917.568678] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dbc3028d-38f8-4498-812b-84f422a4ae14 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.575685] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Uploading image 061f0060-350a-4bf7-a725-58acba59de2b {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 917.601150] env[62753]: DEBUG oslo_vmware.rw_handles [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 917.601150] env[62753]: value = "vm-284620" [ 917.601150] env[62753]: _type = "VirtualMachine" [ 917.601150] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 917.601390] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-af653ebe-ebee-4a99-9d48-256bd8a042b1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.607670] env[62753]: DEBUG oslo_vmware.rw_handles [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lease: (returnval){ [ 917.607670] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52061577-b4ff-cb37-38be-9ff51daec6c5" [ 917.607670] env[62753]: _type = "HttpNfcLease" [ 917.607670] env[62753]: } obtained for exporting VM: (result){ [ 917.607670] env[62753]: value = "vm-284620" [ 917.607670] env[62753]: _type = "VirtualMachine" [ 917.607670] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 917.607951] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the lease: (returnval){ [ 917.607951] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52061577-b4ff-cb37-38be-9ff51daec6c5" [ 917.607951] env[62753]: _type = "HttpNfcLease" [ 917.607951] env[62753]: } to be ready. {{(pid=62753) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 917.616706] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 917.616706] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52061577-b4ff-cb37-38be-9ff51daec6c5" [ 917.616706] env[62753]: _type = "HttpNfcLease" [ 917.616706] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 917.683623] env[62753]: DEBUG nova.network.neutron [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 917.697337] env[62753]: DEBUG oslo_vmware.api [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': task-1332380, 'name': PowerOnVM_Task, 'duration_secs': 0.598785} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.697565] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 917.697767] env[62753]: INFO nova.compute.manager [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Took 6.90 seconds to spawn the instance on the hypervisor. [ 917.697946] env[62753]: DEBUG nova.compute.manager [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 917.698766] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f826474c-5da6-4a9f-b517-f8fb9529c57e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.771198] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6bd8d782-b514-4750-9efa-4696331d2f04 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "49311ef5-733a-4146-87ac-876a121a8d4d" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 26.750s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.786913] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]524e2667-ac62-92ff-b9c4-12e0fdbac39f, 'name': SearchDatastore_Task, 'duration_secs': 0.021397} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 917.788201] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f259b798-5aa1-4f88-8998-c29dc68c9e87 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.793894] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 917.793894] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d1ca7e-39bb-cb52-5fdc-a8172fcef4a2" [ 917.793894] env[62753]: _type = "Task" [ 917.793894] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 917.801796] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d1ca7e-39bb-cb52-5fdc-a8172fcef4a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 917.865167] env[62753]: DEBUG nova.compute.manager [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 917.871093] env[62753]: DEBUG nova.compute.manager [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Received event network-vif-plugged-7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.871342] env[62753]: DEBUG oslo_concurrency.lockutils [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] Acquiring lock "a401909c-5978-492b-a4fa-b17d38a5af44-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 917.871560] env[62753]: DEBUG oslo_concurrency.lockutils [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] Lock "a401909c-5978-492b-a4fa-b17d38a5af44-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 917.871735] env[62753]: DEBUG oslo_concurrency.lockutils [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] Lock "a401909c-5978-492b-a4fa-b17d38a5af44-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 917.871901] env[62753]: DEBUG nova.compute.manager [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] No waiting events found dispatching network-vif-plugged-7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 917.872076] env[62753]: WARNING nova.compute.manager [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Received unexpected event network-vif-plugged-7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36 for instance with vm_state building and task_state spawning. [ 917.872241] env[62753]: DEBUG nova.compute.manager [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Received event network-changed-7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 917.872397] env[62753]: DEBUG nova.compute.manager [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Refreshing instance network info cache due to event network-changed-7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 917.872564] env[62753]: DEBUG oslo_concurrency.lockutils [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] Acquiring lock "refresh_cache-a401909c-5978-492b-a4fa-b17d38a5af44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 917.887841] env[62753]: DEBUG nova.virt.hardware [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 917.888115] env[62753]: DEBUG nova.virt.hardware [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 917.888283] env[62753]: DEBUG nova.virt.hardware [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 917.888493] env[62753]: DEBUG nova.virt.hardware [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 917.888665] env[62753]: DEBUG nova.virt.hardware [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 917.888820] env[62753]: DEBUG nova.virt.hardware [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 917.889301] env[62753]: DEBUG nova.virt.hardware [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 917.889534] env[62753]: DEBUG nova.virt.hardware [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 917.889726] env[62753]: DEBUG nova.virt.hardware [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 917.889899] env[62753]: DEBUG nova.virt.hardware [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 917.890354] env[62753]: DEBUG nova.virt.hardware [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 917.891755] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e857639-5dc4-4c12-b638-d331252a4e64 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 917.901761] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8518caa4-9134-40ff-8e76-d65f6daf8a36 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.069544] env[62753]: DEBUG nova.network.neutron [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Updating instance_info_cache with network_info: [{"id": "7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36", "address": "fa:16:3e:52:63:e3", "network": {"id": "2823647f-d32d-4a1c-9e02-7577016a260e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1594498922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6c8a6bc946a469fa85cc3ab80d7333a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7df5849e-e6", "ovs_interfaceid": "7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 918.119159] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 918.119159] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52061577-b4ff-cb37-38be-9ff51daec6c5" [ 918.119159] env[62753]: _type = "HttpNfcLease" [ 918.119159] env[62753]: } is ready. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 918.119479] env[62753]: DEBUG oslo_vmware.rw_handles [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 918.119479] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52061577-b4ff-cb37-38be-9ff51daec6c5" [ 918.119479] env[62753]: _type = "HttpNfcLease" [ 918.119479] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 918.120409] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d0ef112-f705-48cf-bd89-f52d401ff067 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.137750] env[62753]: DEBUG oslo_vmware.rw_handles [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520cfbbc-db58-305e-9b68-0096aff0aa7d/disk-0.vmdk from lease info. {{(pid=62753) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 918.137962] env[62753]: DEBUG oslo_vmware.rw_handles [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520cfbbc-db58-305e-9b68-0096aff0aa7d/disk-0.vmdk for reading. {{(pid=62753) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 918.195392] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.195600] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.195803] env[62753]: DEBUG nova.network.neutron [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 918.219138] env[62753]: INFO nova.compute.manager [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Took 28.59 seconds to build instance. [ 918.290121] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 3bc05fdb-81be-4764-b9d9-01acd125b020 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 918.290489] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 918.290489] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance c5dae9a3-052b-4f4c-ac79-84aeaded457a actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 918.290662] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 4d20e083-2959-453a-8875-47955bc02613 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 918.290775] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 50e1c1a1-7b8f-49cd-932d-03c920209634 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 918.290981] env[62753]: WARNING nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 33c1de85-aade-4c69-8fdb-7672d8c53300 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 918.291203] env[62753]: WARNING nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 57eef548-a0f1-4f0b-a026-885a10d005c2 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 918.291366] env[62753]: WARNING nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance f03739f4-921f-46e5-b0e7-e9a99d96d621 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 918.291586] env[62753]: WARNING nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 918.291745] env[62753]: WARNING nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 6b9a08b0-3cb3-474a-8193-953e74252a53 is not being actively managed by this compute host but has allocations referencing this compute host: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. Skipping heal of allocation because we do not know what to do. [ 918.291831] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance c73b1ae0-dc98-47f7-babf-e96169384785 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 918.292073] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 77cdd901-cc96-4a6f-a696-a65f54a96b1a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 918.292415] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 918.292415] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance b5674964-1928-4ecb-b1fd-8f60a94b4270 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 918.292415] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance a401909c-5978-492b-a4fa-b17d38a5af44 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 918.292634] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance b45bfcb4-b2cd-468c-9595-62c155d520fb actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 918.293010] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 918.293010] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2624MB phys_disk=200GB used_disk=10GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 918.296456] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-14e5d932-5d9e-4c6e-9acc-9c875abe3508 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.298399] env[62753]: DEBUG nova.network.neutron [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Successfully updated port: 91add90d-924f-43de-bead-d4f33e2af4ad {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 918.310015] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d1ca7e-39bb-cb52-5fdc-a8172fcef4a2, 'name': SearchDatastore_Task, 'duration_secs': 0.012626} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.312533] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.312533] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] b5674964-1928-4ecb-b1fd-8f60a94b4270/b5674964-1928-4ecb-b1fd-8f60a94b4270.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 918.312533] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-7e62deba-573f-4ed4-ace8-937eca68d1fe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.318819] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 918.318819] env[62753]: value = "task-1332382" [ 918.318819] env[62753]: _type = "Task" [ 918.318819] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.326983] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332382, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.351181] env[62753]: DEBUG nova.compute.manager [req-27fbf4dd-c2da-4234-91a7-750dc548e684 req-b07b33bc-6612-449a-913d-66956e7ba67e service nova] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Received event network-vif-plugged-91add90d-924f-43de-bead-d4f33e2af4ad {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 918.351181] env[62753]: DEBUG oslo_concurrency.lockutils [req-27fbf4dd-c2da-4234-91a7-750dc548e684 req-b07b33bc-6612-449a-913d-66956e7ba67e service nova] Acquiring lock "b45bfcb4-b2cd-468c-9595-62c155d520fb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 918.351181] env[62753]: DEBUG oslo_concurrency.lockutils [req-27fbf4dd-c2da-4234-91a7-750dc548e684 req-b07b33bc-6612-449a-913d-66956e7ba67e service nova] Lock "b45bfcb4-b2cd-468c-9595-62c155d520fb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 918.351181] env[62753]: DEBUG oslo_concurrency.lockutils [req-27fbf4dd-c2da-4234-91a7-750dc548e684 req-b07b33bc-6612-449a-913d-66956e7ba67e service nova] Lock "b45bfcb4-b2cd-468c-9595-62c155d520fb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.351415] env[62753]: DEBUG nova.compute.manager [req-27fbf4dd-c2da-4234-91a7-750dc548e684 req-b07b33bc-6612-449a-913d-66956e7ba67e service nova] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] No waiting events found dispatching network-vif-plugged-91add90d-924f-43de-bead-d4f33e2af4ad {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 918.351415] env[62753]: WARNING nova.compute.manager [req-27fbf4dd-c2da-4234-91a7-750dc548e684 req-b07b33bc-6612-449a-913d-66956e7ba67e service nova] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Received unexpected event network-vif-plugged-91add90d-924f-43de-bead-d4f33e2af4ad for instance with vm_state building and task_state spawning. [ 918.562019] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c39ea5f7-4957-4cc2-9bf3-cf303534db33 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.575785] env[62753]: DEBUG oslo_concurrency.lockutils [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "refresh_cache-a401909c-5978-492b-a4fa-b17d38a5af44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 918.576276] env[62753]: DEBUG nova.compute.manager [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Instance network_info: |[{"id": "7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36", "address": "fa:16:3e:52:63:e3", "network": {"id": "2823647f-d32d-4a1c-9e02-7577016a260e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1594498922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6c8a6bc946a469fa85cc3ab80d7333a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7df5849e-e6", "ovs_interfaceid": "7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 918.578078] env[62753]: DEBUG oslo_concurrency.lockutils [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] Acquired lock "refresh_cache-a401909c-5978-492b-a4fa-b17d38a5af44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.578078] env[62753]: DEBUG nova.network.neutron [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Refreshing network info cache for port 7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 918.579341] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:52:63:e3', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9d39252e-42ef-4252-98d3-62af5a0d109d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 918.597249] env[62753]: DEBUG oslo.service.loopingcall [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 918.598771] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5feb794-63d2-438c-86c1-22574fe4559f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.607528] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 918.608286] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-29994fd7-d552-4303-b4b6-a7eb76469a91 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.661749] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-644e0d6d-b23d-4def-b188-2ed1290c7961 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.665258] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 918.665258] env[62753]: value = "task-1332383" [ 918.665258] env[62753]: _type = "Task" [ 918.665258] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.672959] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3728991-b5dc-40b3-a069-b59a164b8315 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.683048] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332383, 'name': CreateVM_Task} progress is 15%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 918.695757] env[62753]: DEBUG nova.compute.provider_tree [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 918.721237] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c76865c2-416f-4cb9-8059-bb4c19b8327c tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Lock "d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 53.732s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 918.805036] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "refresh_cache-b45bfcb4-b2cd-468c-9595-62c155d520fb" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 918.806110] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquired lock "refresh_cache-b45bfcb4-b2cd-468c-9595-62c155d520fb" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 918.806110] env[62753]: DEBUG nova.network.neutron [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 918.829796] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332382, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.473003} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 918.833637] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] b5674964-1928-4ecb-b1fd-8f60a94b4270/b5674964-1928-4ecb-b1fd-8f60a94b4270.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 918.834204] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 918.834688] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-87d22157-1049-46f1-8a98-2a62fad550e6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 918.842512] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 918.842512] env[62753]: value = "task-1332384" [ 918.842512] env[62753]: _type = "Task" [ 918.842512] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 918.861869] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332384, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.062884] env[62753]: DEBUG nova.network.neutron [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Updated VIF entry in instance network info cache for port 7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 919.063339] env[62753]: DEBUG nova.network.neutron [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Updating instance_info_cache with network_info: [{"id": "7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36", "address": "fa:16:3e:52:63:e3", "network": {"id": "2823647f-d32d-4a1c-9e02-7577016a260e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1594498922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6c8a6bc946a469fa85cc3ab80d7333a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7df5849e-e6", "ovs_interfaceid": "7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.178388] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332383, 'name': CreateVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.202434] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 919.219570] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "3bc05fdb-81be-4764-b9d9-01acd125b020" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.220089] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "3bc05fdb-81be-4764-b9d9-01acd125b020" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.220340] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "3bc05fdb-81be-4764-b9d9-01acd125b020-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 919.220456] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "3bc05fdb-81be-4764-b9d9-01acd125b020-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.220817] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "3bc05fdb-81be-4764-b9d9-01acd125b020-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.222936] env[62753]: INFO nova.compute.manager [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Terminating instance [ 919.225833] env[62753]: DEBUG nova.compute.manager [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 919.226140] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 919.229495] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce78261e-bf07-43a7-af62-707b97aaf9ef {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.237568] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 919.237838] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1f76ba19-34f6-43e1-91f7-e17687675495 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.245593] env[62753]: DEBUG oslo_vmware.api [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 919.245593] env[62753]: value = "task-1332385" [ 919.245593] env[62753]: _type = "Task" [ 919.245593] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.255779] env[62753]: DEBUG oslo_vmware.api [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332385, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.358772] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332384, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074419} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 919.359904] env[62753]: DEBUG nova.network.neutron [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 919.362041] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 919.363432] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33cb0bac-f77b-4daa-9532-44c38dfaa2c6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.390626] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Reconfiguring VM instance instance-0000004b to attach disk [datastore1] b5674964-1928-4ecb-b1fd-8f60a94b4270/b5674964-1928-4ecb-b1fd-8f60a94b4270.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 919.390626] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b5576eb4-6363-4b8b-8396-218dc5ff3714 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 919.413817] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 919.413817] env[62753]: value = "task-1332386" [ 919.413817] env[62753]: _type = "Task" [ 919.413817] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 919.422138] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332386, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.571025] env[62753]: DEBUG oslo_concurrency.lockutils [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] Releasing lock "refresh_cache-a401909c-5978-492b-a4fa-b17d38a5af44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 919.571025] env[62753]: DEBUG nova.compute.manager [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Received event network-vif-deleted-9713beba-767d-408c-bd7d-bef58dc3c405 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.571025] env[62753]: INFO nova.compute.manager [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Neutron deleted interface 9713beba-767d-408c-bd7d-bef58dc3c405; detaching it from the instance and deleting it from the info cache [ 919.571025] env[62753]: DEBUG nova.network.neutron [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Updating instance_info_cache with network_info: [{"id": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "address": "fa:16:3e:1d:84:1b", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8e502a-7a", "ovs_interfaceid": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "43ffe16d-d249-47ae-842c-564f9274c653", "address": "fa:16:3e:a8:c4:dc", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap43ffe16d-d2", "ovs_interfaceid": "43ffe16d-d249-47ae-842c-564f9274c653", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.573227] env[62753]: INFO nova.network.neutron [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Port 9713beba-767d-408c-bd7d-bef58dc3c405 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 919.573511] env[62753]: INFO nova.network.neutron [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Port 43ffe16d-d249-47ae-842c-564f9274c653 from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 919.573930] env[62753]: DEBUG nova.network.neutron [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Updating instance_info_cache with network_info: [{"id": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "address": "fa:16:3e:1d:84:1b", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.207", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7e8e502a-7a", "ovs_interfaceid": "7e8e502a-7a64-4dfd-8f9d-18965ac19b6b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.679310] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332383, 'name': CreateVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.708098] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 919.708348] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.467s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 919.708635] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 16.883s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 919.708817] env[62753]: DEBUG nova.objects.instance [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62753) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 919.760438] env[62753]: DEBUG oslo_vmware.api [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332385, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.766154] env[62753]: DEBUG nova.network.neutron [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Updating instance_info_cache with network_info: [{"id": "91add90d-924f-43de-bead-d4f33e2af4ad", "address": "fa:16:3e:88:31:f0", "network": {"id": "12446a5e-7329-4f8f-aa28-a8b9d78e7573", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1766107072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "371a5e8b16be4cffbdeb2139757684fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7c7f16b-a2e3-4d1f-9b7e-a44a9ebb589c", "external-id": "nsx-vlan-transportzone-772", "segmentation_id": 772, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91add90d-92", "ovs_interfaceid": "91add90d-924f-43de-bead-d4f33e2af4ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 919.923759] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332386, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 919.977164] env[62753]: DEBUG nova.compute.manager [req-6f0b27ad-4c21-4be7-bee9-1fe84c308306 req-347aa69f-ea60-4a54-86b2-b112439cc86d service nova] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Received event network-changed-4037b488-2f26-40ee-b044-3b4fec79a454 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 919.977432] env[62753]: DEBUG nova.compute.manager [req-6f0b27ad-4c21-4be7-bee9-1fe84c308306 req-347aa69f-ea60-4a54-86b2-b112439cc86d service nova] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Refreshing instance network info cache due to event network-changed-4037b488-2f26-40ee-b044-3b4fec79a454. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 919.977683] env[62753]: DEBUG oslo_concurrency.lockutils [req-6f0b27ad-4c21-4be7-bee9-1fe84c308306 req-347aa69f-ea60-4a54-86b2-b112439cc86d service nova] Acquiring lock "refresh_cache-77cdd901-cc96-4a6f-a696-a65f54a96b1a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 919.977834] env[62753]: DEBUG oslo_concurrency.lockutils [req-6f0b27ad-4c21-4be7-bee9-1fe84c308306 req-347aa69f-ea60-4a54-86b2-b112439cc86d service nova] Acquired lock "refresh_cache-77cdd901-cc96-4a6f-a696-a65f54a96b1a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 919.978042] env[62753]: DEBUG nova.network.neutron [req-6f0b27ad-4c21-4be7-bee9-1fe84c308306 req-347aa69f-ea60-4a54-86b2-b112439cc86d service nova] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Refreshing network info cache for port 4037b488-2f26-40ee-b044-3b4fec79a454 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 920.072815] env[62753]: DEBUG oslo_concurrency.lockutils [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] Acquiring lock "3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.076590] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "refresh_cache-3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.180020] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332383, 'name': CreateVM_Task, 'duration_secs': 1.490101} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.180020] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 920.180020] env[62753]: DEBUG oslo_concurrency.lockutils [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.180020] env[62753]: DEBUG oslo_concurrency.lockutils [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.180020] env[62753]: DEBUG oslo_concurrency.lockutils [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 920.180020] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7a1f9ef-3106-4ca2-9bf9-d8a1f9484ca3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.185696] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 920.185696] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e47ed0-dd92-596d-49e0-f9916565b2ab" [ 920.185696] env[62753]: _type = "Task" [ 920.185696] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.195987] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e47ed0-dd92-596d-49e0-f9916565b2ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.260497] env[62753]: DEBUG oslo_vmware.api [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332385, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.269586] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Releasing lock "refresh_cache-b45bfcb4-b2cd-468c-9595-62c155d520fb" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.269908] env[62753]: DEBUG nova.compute.manager [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Instance network_info: |[{"id": "91add90d-924f-43de-bead-d4f33e2af4ad", "address": "fa:16:3e:88:31:f0", "network": {"id": "12446a5e-7329-4f8f-aa28-a8b9d78e7573", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1766107072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "371a5e8b16be4cffbdeb2139757684fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7c7f16b-a2e3-4d1f-9b7e-a44a9ebb589c", "external-id": "nsx-vlan-transportzone-772", "segmentation_id": 772, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91add90d-92", "ovs_interfaceid": "91add90d-924f-43de-bead-d4f33e2af4ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 920.270682] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:31:f0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c7c7f16b-a2e3-4d1f-9b7e-a44a9ebb589c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '91add90d-924f-43de-bead-d4f33e2af4ad', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 920.279943] env[62753]: DEBUG oslo.service.loopingcall [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 920.280292] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 920.280664] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3debdca7-19aa-4101-9c2d-19b93ea44fe7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.302281] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 920.302281] env[62753]: value = "task-1332387" [ 920.302281] env[62753]: _type = "Task" [ 920.302281] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.314075] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332387, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.387922] env[62753]: DEBUG nova.compute.manager [req-feb22dfa-6905-4209-8d49-3af56f4b4660 req-e5a1c9c4-d486-47a1-9400-dedb5626b69a service nova] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Received event network-changed-91add90d-924f-43de-bead-d4f33e2af4ad {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 920.388248] env[62753]: DEBUG nova.compute.manager [req-feb22dfa-6905-4209-8d49-3af56f4b4660 req-e5a1c9c4-d486-47a1-9400-dedb5626b69a service nova] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Refreshing instance network info cache due to event network-changed-91add90d-924f-43de-bead-d4f33e2af4ad. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 920.388607] env[62753]: DEBUG oslo_concurrency.lockutils [req-feb22dfa-6905-4209-8d49-3af56f4b4660 req-e5a1c9c4-d486-47a1-9400-dedb5626b69a service nova] Acquiring lock "refresh_cache-b45bfcb4-b2cd-468c-9595-62c155d520fb" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.388936] env[62753]: DEBUG oslo_concurrency.lockutils [req-feb22dfa-6905-4209-8d49-3af56f4b4660 req-e5a1c9c4-d486-47a1-9400-dedb5626b69a service nova] Acquired lock "refresh_cache-b45bfcb4-b2cd-468c-9595-62c155d520fb" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.389099] env[62753]: DEBUG nova.network.neutron [req-feb22dfa-6905-4209-8d49-3af56f4b4660 req-e5a1c9c4-d486-47a1-9400-dedb5626b69a service nova] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Refreshing network info cache for port 91add90d-924f-43de-bead-d4f33e2af4ad {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 920.426528] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332386, 'name': ReconfigVM_Task, 'duration_secs': 0.613897} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.426850] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Reconfigured VM instance instance-0000004b to attach disk [datastore1] b5674964-1928-4ecb-b1fd-8f60a94b4270/b5674964-1928-4ecb-b1fd-8f60a94b4270.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 920.427792] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-3306f464-9ad8-4ec4-b902-471bc1a5a947 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.435397] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 920.435397] env[62753]: value = "task-1332388" [ 920.435397] env[62753]: _type = "Task" [ 920.435397] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.446263] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332388, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.580564] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd92d214-50f9-4cbd-85e5-f4d495d3e631 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "interface-3bc05fdb-81be-4764-b9d9-01acd125b020-9713beba-767d-408c-bd7d-bef58dc3c405" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.395s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.708099] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 920.713055] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e47ed0-dd92-596d-49e0-f9916565b2ab, 'name': SearchDatastore_Task, 'duration_secs': 0.069405} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.713055] env[62753]: DEBUG oslo_concurrency.lockutils [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.713055] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 920.713055] env[62753]: DEBUG oslo_concurrency.lockutils [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.713055] env[62753]: DEBUG oslo_concurrency.lockutils [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.713055] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 920.713055] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9b71f418-9a13-4864-ae66-e515ba7cb334 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.717347] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a46ea38-2fd5-4fd9-ae5f-75282c530ed1 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.718743] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 17.297s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.719118] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.721506] env[62753]: DEBUG oslo_concurrency.lockutils [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 15.542s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.721846] env[62753]: DEBUG oslo_concurrency.lockutils [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.724198] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 12.619s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.724744] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.726979] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 11.433s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.727370] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.729466] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.022s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 920.732828] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 920.734575] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 920.734927] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 920.736117] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-947490f8-75d6-4f3c-8e53-47f0a236bcf2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.741329] env[62753]: DEBUG nova.network.neutron [req-6f0b27ad-4c21-4be7-bee9-1fe84c308306 req-347aa69f-ea60-4a54-86b2-b112439cc86d service nova] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Updated VIF entry in instance network info cache for port 4037b488-2f26-40ee-b044-3b4fec79a454. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 920.742328] env[62753]: DEBUG nova.network.neutron [req-6f0b27ad-4c21-4be7-bee9-1fe84c308306 req-347aa69f-ea60-4a54-86b2-b112439cc86d service nova] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Updating instance_info_cache with network_info: [{"id": "4037b488-2f26-40ee-b044-3b4fec79a454", "address": "fa:16:3e:6f:b1:75", "network": {"id": "56efb90b-64b4-484e-8062-eef8d39b1caf", "bridge": "br-int", "label": "tempest-AttachVolumeTestJSON-1294891206-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.170", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "909db6176c8f475caf2562a593bc61a9", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8480e3f5-68bd-4c27-ae1f-7c994a8202b1", "external-id": "nsx-vlan-transportzone-628", "segmentation_id": 628, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4037b488-2f", "ovs_interfaceid": "4037b488-2f26-40ee-b044-3b4fec79a454", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 920.747110] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 920.747110] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529553bf-e385-d8a1-f5fe-4be697e0d481" [ 920.747110] env[62753]: _type = "Task" [ 920.747110] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.763662] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529553bf-e385-d8a1-f5fe-4be697e0d481, 'name': SearchDatastore_Task, 'duration_secs': 0.015785} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.768563] env[62753]: DEBUG oslo_vmware.api [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332385, 'name': PowerOffVM_Task, 'duration_secs': 1.066917} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.769745] env[62753]: INFO nova.scheduler.client.report [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Deleted allocations for instance 6b9a08b0-3cb3-474a-8193-953e74252a53 [ 920.770789] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3cd02b35-b291-49c0-837c-82fff612468e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.778289] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 920.778444] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 920.779501] env[62753]: INFO nova.scheduler.client.report [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Deleted allocations for instance 33c1de85-aade-4c69-8fdb-7672d8c53300 [ 920.783276] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9b7cc511-55a1-450a-aad1-800f0b0bdaaf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.789797] env[62753]: INFO nova.scheduler.client.report [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Deleted allocations for instance 57eef548-a0f1-4f0b-a026-885a10d005c2 [ 920.793647] env[62753]: INFO nova.scheduler.client.report [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Deleted allocations for instance 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66 [ 920.798541] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 920.798541] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52665f31-0a0b-32fc-da09-57959056d63c" [ 920.798541] env[62753]: _type = "Task" [ 920.798541] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.813951] env[62753]: INFO nova.scheduler.client.report [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Deleted allocations for instance f03739f4-921f-46e5-b0e7-e9a99d96d621 [ 920.821622] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52665f31-0a0b-32fc-da09-57959056d63c, 'name': SearchDatastore_Task, 'duration_secs': 0.013693} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.824581] env[62753]: DEBUG oslo_concurrency.lockutils [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.824850] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] a401909c-5978-492b-a4fa-b17d38a5af44/a401909c-5978-492b-a4fa-b17d38a5af44.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 920.829586] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b6eaabe9-4fe7-476c-87e7-5dd5571a8988 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.832392] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332387, 'name': CreateVM_Task, 'duration_secs': 0.417322} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.832708] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 920.833808] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.834014] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.834398] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 920.835194] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7d6af908-03cb-47fb-a93b-157c5fb61389 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.838242] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 920.838242] env[62753]: value = "task-1332390" [ 920.838242] env[62753]: _type = "Task" [ 920.838242] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.843686] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 920.843686] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a53bf4-58c2-3964-6ed0-b08a73b52225" [ 920.843686] env[62753]: _type = "Task" [ 920.843686] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.852209] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332390, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.857787] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a53bf4-58c2-3964-6ed0-b08a73b52225, 'name': SearchDatastore_Task, 'duration_secs': 0.010079} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.858302] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 920.858650] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 920.858953] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 920.859164] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 920.859375] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 920.859680] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-129f4cc2-d901-433f-b559-1dce76fa8339 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.869700] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 920.869936] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 920.870901] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a91f69d9-fd88-4d39-b7a8-44bdaa1b1fc0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.878057] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 920.878057] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5299a60e-da51-fd07-bff5-7aaf0973129a" [ 920.878057] env[62753]: _type = "Task" [ 920.878057] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.888778] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5299a60e-da51-fd07-bff5-7aaf0973129a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.899774] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 920.900050] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 920.900272] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Deleting the datastore file [datastore2] 3bc05fdb-81be-4764-b9d9-01acd125b020 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 920.900931] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-5d8f6c2f-fbab-44e3-9e03-fee33a9ab32d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.908076] env[62753]: DEBUG oslo_vmware.api [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 920.908076] env[62753]: value = "task-1332391" [ 920.908076] env[62753]: _type = "Task" [ 920.908076] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.917656] env[62753]: DEBUG oslo_vmware.api [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332391, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 920.947229] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332388, 'name': Rename_Task, 'duration_secs': 0.220402} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 920.947652] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 920.947980] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-eadeaabd-723f-455b-8dcd-dac0172111d4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 920.954412] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 920.954412] env[62753]: value = "task-1332392" [ 920.954412] env[62753]: _type = "Task" [ 920.954412] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 920.963701] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332392, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.221206] env[62753]: DEBUG nova.network.neutron [req-feb22dfa-6905-4209-8d49-3af56f4b4660 req-e5a1c9c4-d486-47a1-9400-dedb5626b69a service nova] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Updated VIF entry in instance network info cache for port 91add90d-924f-43de-bead-d4f33e2af4ad. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 921.221206] env[62753]: DEBUG nova.network.neutron [req-feb22dfa-6905-4209-8d49-3af56f4b4660 req-e5a1c9c4-d486-47a1-9400-dedb5626b69a service nova] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Updating instance_info_cache with network_info: [{"id": "91add90d-924f-43de-bead-d4f33e2af4ad", "address": "fa:16:3e:88:31:f0", "network": {"id": "12446a5e-7329-4f8f-aa28-a8b9d78e7573", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1766107072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "371a5e8b16be4cffbdeb2139757684fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7c7f16b-a2e3-4d1f-9b7e-a44a9ebb589c", "external-id": "nsx-vlan-transportzone-772", "segmentation_id": 772, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap91add90d-92", "ovs_interfaceid": "91add90d-924f-43de-bead-d4f33e2af4ad", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 921.222646] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 921.224272] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Starting heal instance info cache {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 921.247089] env[62753]: DEBUG oslo_concurrency.lockutils [req-6f0b27ad-4c21-4be7-bee9-1fe84c308306 req-347aa69f-ea60-4a54-86b2-b112439cc86d service nova] Releasing lock "refresh_cache-77cdd901-cc96-4a6f-a696-a65f54a96b1a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.247229] env[62753]: DEBUG nova.compute.manager [req-6f0b27ad-4c21-4be7-bee9-1fe84c308306 req-347aa69f-ea60-4a54-86b2-b112439cc86d service nova] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Received event network-changed-ef993aa6-10f5-4de9-b652-426dbaac5112 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.247497] env[62753]: DEBUG nova.compute.manager [req-6f0b27ad-4c21-4be7-bee9-1fe84c308306 req-347aa69f-ea60-4a54-86b2-b112439cc86d service nova] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Refreshing instance network info cache due to event network-changed-ef993aa6-10f5-4de9-b652-426dbaac5112. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 921.247779] env[62753]: DEBUG oslo_concurrency.lockutils [req-6f0b27ad-4c21-4be7-bee9-1fe84c308306 req-347aa69f-ea60-4a54-86b2-b112439cc86d service nova] Acquiring lock "refresh_cache-d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 921.247967] env[62753]: DEBUG oslo_concurrency.lockutils [req-6f0b27ad-4c21-4be7-bee9-1fe84c308306 req-347aa69f-ea60-4a54-86b2-b112439cc86d service nova] Acquired lock "refresh_cache-d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 921.250128] env[62753]: DEBUG nova.network.neutron [req-6f0b27ad-4c21-4be7-bee9-1fe84c308306 req-347aa69f-ea60-4a54-86b2-b112439cc86d service nova] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Refreshing network info cache for port ef993aa6-10f5-4de9-b652-426dbaac5112 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 921.289987] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8a6ebea8-e867-4b8a-b1fc-a9d92cb42a66 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "6b9a08b0-3cb3-474a-8193-953e74252a53" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.943s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.299034] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e81b89f7-f5de-427e-a195-200131715b27 tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "33c1de85-aade-4c69-8fdb-7672d8c53300" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 16.925s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.307195] env[62753]: DEBUG oslo_concurrency.lockutils [None req-70f5f40c-1371-4146-860c-faa8605b8e6d tempest-ServerAddressesNegativeTestJSON-839010597 tempest-ServerAddressesNegativeTestJSON-839010597-project-member] Lock "9a8a148c-057d-4c91-8b7c-4e05e6ac0f66" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 19.040s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.309814] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7c00d66c-a01d-4258-9111-74fe01696176 tempest-ServerMetadataNegativeTestJSON-475200133 tempest-ServerMetadataNegativeTestJSON-475200133-project-member] Lock "57eef548-a0f1-4f0b-a026-885a10d005c2" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 21.329s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.332998] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5217710a-bc04-424f-8fec-7a7baa281664 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "f03739f4-921f-46e5-b0e7-e9a99d96d621" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.244s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 921.357745] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332390, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.392123] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5299a60e-da51-fd07-bff5-7aaf0973129a, 'name': SearchDatastore_Task, 'duration_secs': 0.010077} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.393221] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ef8b283a-08b5-4922-9c03-477d4d54e7f0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.399820] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 921.399820] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52efe742-d4a4-1d75-f277-c2e64248a3d1" [ 921.399820] env[62753]: _type = "Task" [ 921.399820] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.409705] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52efe742-d4a4-1d75-f277-c2e64248a3d1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.420586] env[62753]: DEBUG oslo_vmware.api [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332391, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.352828} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.421166] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 921.424038] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 921.424038] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 921.424038] env[62753]: INFO nova.compute.manager [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Took 2.20 seconds to destroy the instance on the hypervisor. [ 921.424038] env[62753]: DEBUG oslo.service.loopingcall [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 921.424038] env[62753]: DEBUG nova.compute.manager [-] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 921.424038] env[62753]: DEBUG nova.network.neutron [-] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 921.471149] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332392, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.487060] env[62753]: DEBUG neutronclient.v2_0.client [-] Error message: {"NeutronError": {"type": "PortNotFound", "message": "Port 43ffe16d-d249-47ae-842c-564f9274c653 could not be found.", "detail": ""}} {{(pid=62753) _handle_fault_response /opt/stack/data/venv/lib/python3.10/site-packages/neutronclient/v2_0/client.py:262}} [ 921.487060] env[62753]: DEBUG nova.network.neutron [-] Unable to show port 43ffe16d-d249-47ae-842c-564f9274c653 as it no longer exists. {{(pid=62753) _unbind_ports /opt/stack/nova/nova/network/neutron.py:666}} [ 921.727713] env[62753]: DEBUG oslo_concurrency.lockutils [req-feb22dfa-6905-4209-8d49-3af56f4b4660 req-e5a1c9c4-d486-47a1-9400-dedb5626b69a service nova] Releasing lock "refresh_cache-b45bfcb4-b2cd-468c-9595-62c155d520fb" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.727713] env[62753]: DEBUG nova.compute.manager [req-feb22dfa-6905-4209-8d49-3af56f4b4660 req-e5a1c9c4-d486-47a1-9400-dedb5626b69a service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Received event network-vif-deleted-43ffe16d-d249-47ae-842c-564f9274c653 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 921.850659] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332390, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.912298] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52efe742-d4a4-1d75-f277-c2e64248a3d1, 'name': SearchDatastore_Task, 'duration_secs': 0.034583} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.912996] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 921.913589] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] b45bfcb4-b2cd-468c-9595-62c155d520fb/b45bfcb4-b2cd-468c-9595-62c155d520fb.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 921.914831] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-754425d0-4c1a-4880-9150-b3dde1aad61b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 921.925094] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 921.925094] env[62753]: value = "task-1332393" [ 921.925094] env[62753]: _type = "Task" [ 921.925094] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 921.933949] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332393, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 921.964613] env[62753]: DEBUG oslo_vmware.api [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332392, 'name': PowerOnVM_Task, 'duration_secs': 0.73292} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 921.965823] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 921.966112] env[62753]: INFO nova.compute.manager [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Took 8.85 seconds to spawn the instance on the hypervisor. [ 921.966352] env[62753]: DEBUG nova.compute.manager [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 921.967252] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bac245f3-13a9-4d6b-abc7-5d3e0abb8f03 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.179035] env[62753]: DEBUG oslo_concurrency.lockutils [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquiring lock "f8470f38-0150-41c6-a3ee-aabb5e68d98b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.179391] env[62753]: DEBUG oslo_concurrency.lockutils [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "f8470f38-0150-41c6-a3ee-aabb5e68d98b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.194258] env[62753]: DEBUG nova.network.neutron [req-6f0b27ad-4c21-4be7-bee9-1fe84c308306 req-347aa69f-ea60-4a54-86b2-b112439cc86d service nova] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Updated VIF entry in instance network info cache for port ef993aa6-10f5-4de9-b652-426dbaac5112. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 922.196893] env[62753]: DEBUG nova.network.neutron [req-6f0b27ad-4c21-4be7-bee9-1fe84c308306 req-347aa69f-ea60-4a54-86b2-b112439cc86d service nova] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Updating instance_info_cache with network_info: [{"id": "ef993aa6-10f5-4de9-b652-426dbaac5112", "address": "fa:16:3e:37:0d:b5", "network": {"id": "978c6598-cb9f-4d90-9ccb-1de3097a91a6", "bridge": "br-int", "label": "tempest-ServersV294TestFqdnHostnames-110846568-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.138", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "c2dc6fe36ae84435a8e09cdf56340c58", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "1323cb03-8367-485a-962e-131af8eba474", "external-id": "nsx-vlan-transportzone-41", "segmentation_id": 41, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapef993aa6-10", "ovs_interfaceid": "ef993aa6-10f5-4de9-b652-426dbaac5112", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.207038] env[62753]: DEBUG oslo_concurrency.lockutils [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "50e1c1a1-7b8f-49cd-932d-03c920209634" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.207038] env[62753]: DEBUG oslo_concurrency.lockutils [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "50e1c1a1-7b8f-49cd-932d-03c920209634" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.207038] env[62753]: DEBUG oslo_concurrency.lockutils [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "50e1c1a1-7b8f-49cd-932d-03c920209634-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.207038] env[62753]: DEBUG oslo_concurrency.lockutils [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "50e1c1a1-7b8f-49cd-932d-03c920209634-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.207038] env[62753]: DEBUG oslo_concurrency.lockutils [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "50e1c1a1-7b8f-49cd-932d-03c920209634-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 922.208261] env[62753]: INFO nova.compute.manager [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Terminating instance [ 922.210845] env[62753]: DEBUG oslo_concurrency.lockutils [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "refresh_cache-50e1c1a1-7b8f-49cd-932d-03c920209634" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 922.211190] env[62753]: DEBUG oslo_concurrency.lockutils [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquired lock "refresh_cache-50e1c1a1-7b8f-49cd-932d-03c920209634" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 922.211389] env[62753]: DEBUG nova.network.neutron [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 922.231322] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Didn't find any instances for network info cache update. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 922.231322] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 922.253048] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "8066aeb2-7be4-46e3-aba6-124fef342a49" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 922.253048] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "8066aeb2-7be4-46e3-aba6-124fef342a49" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 922.352330] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332390, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.421056] env[62753]: DEBUG nova.compute.manager [req-2096ddee-28ce-441c-aae3-528da74d253f req-f39b4fa8-a2a9-40f3-b04e-f6f0aa75be48 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Received event network-vif-deleted-7e8e502a-7a64-4dfd-8f9d-18965ac19b6b {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 922.421056] env[62753]: INFO nova.compute.manager [req-2096ddee-28ce-441c-aae3-528da74d253f req-f39b4fa8-a2a9-40f3-b04e-f6f0aa75be48 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Neutron deleted interface 7e8e502a-7a64-4dfd-8f9d-18965ac19b6b; detaching it from the instance and deleting it from the info cache [ 922.421056] env[62753]: DEBUG nova.network.neutron [req-2096ddee-28ce-441c-aae3-528da74d253f req-f39b4fa8-a2a9-40f3-b04e-f6f0aa75be48 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.440962] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332393, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.492514] env[62753]: INFO nova.compute.manager [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Took 32.18 seconds to build instance. [ 922.681860] env[62753]: DEBUG nova.compute.manager [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 922.699473] env[62753]: DEBUG oslo_concurrency.lockutils [req-6f0b27ad-4c21-4be7-bee9-1fe84c308306 req-347aa69f-ea60-4a54-86b2-b112439cc86d service nova] Releasing lock "refresh_cache-d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 922.750998] env[62753]: DEBUG nova.network.neutron [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 922.756887] env[62753]: DEBUG nova.compute.manager [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 922.769172] env[62753]: DEBUG nova.network.neutron [-] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.855727] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332390, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.631253} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 922.856281] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] a401909c-5978-492b-a4fa-b17d38a5af44/a401909c-5978-492b-a4fa-b17d38a5af44.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 922.856513] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 922.856792] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-51d68ce9-4e65-4c4b-a6be-f0beccc59170 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.867153] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 922.867153] env[62753]: value = "task-1332394" [ 922.867153] env[62753]: _type = "Task" [ 922.867153] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 922.884291] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332394, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.924917] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-eecabf4a-2c64-47b7-b986-2a355ad829dc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.936375] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c1f982d-0c2f-438a-a58e-510a91e27df4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 922.953491] env[62753]: DEBUG nova.network.neutron [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 922.963551] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332393, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 922.991644] env[62753]: DEBUG nova.compute.manager [req-2096ddee-28ce-441c-aae3-528da74d253f req-f39b4fa8-a2a9-40f3-b04e-f6f0aa75be48 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Detach interface failed, port_id=7e8e502a-7a64-4dfd-8f9d-18965ac19b6b, reason: Instance 3bc05fdb-81be-4764-b9d9-01acd125b020 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 922.998967] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ddf1f4a6-6301-4ae4-bf0a-2c74fe0ad73c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "b5674964-1928-4ecb-b1fd-8f60a94b4270" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 48.963s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 923.215827] env[62753]: DEBUG oslo_concurrency.lockutils [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.215827] env[62753]: DEBUG oslo_concurrency.lockutils [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 923.215827] env[62753]: INFO nova.compute.claims [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 923.274674] env[62753]: INFO nova.compute.manager [-] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Took 1.85 seconds to deallocate network for instance. [ 923.292126] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.380331] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332394, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.120876} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.384023] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 923.384023] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f570b690-b216-40b8-88e2-06b874e33491 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.412269] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] a401909c-5978-492b-a4fa-b17d38a5af44/a401909c-5978-492b-a4fa-b17d38a5af44.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 923.412269] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4bcf02d2-cc64-41a3-ac8d-aa0abc4d27a5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.440728] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332393, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.181031} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.443949] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] b45bfcb4-b2cd-468c-9595-62c155d520fb/b45bfcb4-b2cd-468c-9595-62c155d520fb.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 923.443949] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 923.443949] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 923.443949] env[62753]: value = "task-1332395" [ 923.443949] env[62753]: _type = "Task" [ 923.443949] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.443949] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-06368bab-0041-4831-b03f-1d8341b67b7c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.456401] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332395, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.457926] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 923.457926] env[62753]: value = "task-1332396" [ 923.457926] env[62753]: _type = "Task" [ 923.457926] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.465208] env[62753]: DEBUG oslo_concurrency.lockutils [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Releasing lock "refresh_cache-50e1c1a1-7b8f-49cd-932d-03c920209634" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 923.465648] env[62753]: DEBUG nova.compute.manager [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 923.465848] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 923.470584] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3c1fb5b-52bf-410a-9f09-819405e86b32 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.473850] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332396, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.484342] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 923.484342] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6542a4ad-022a-4b05-9085-1f3cefb730b2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 923.493996] env[62753]: DEBUG oslo_vmware.api [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 923.493996] env[62753]: value = "task-1332397" [ 923.493996] env[62753]: _type = "Task" [ 923.493996] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 923.505061] env[62753]: DEBUG oslo_vmware.api [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332397, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.785944] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 923.962302] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332395, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 923.972058] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332396, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.086823} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 923.972655] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 923.977734] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c9a23a4-05ad-4bec-9987-1180a9cf006e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.007201] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Reconfiguring VM instance instance-0000004d to attach disk [datastore1] b45bfcb4-b2cd-468c-9595-62c155d520fb/b45bfcb4-b2cd-468c-9595-62c155d520fb.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 924.010253] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-36b57537-ef42-4ebc-9683-d821c8619c9d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.041688] env[62753]: DEBUG oslo_vmware.api [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332397, 'name': PowerOffVM_Task, 'duration_secs': 0.238533} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.043015] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 924.043292] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 924.043646] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 924.043646] env[62753]: value = "task-1332398" [ 924.043646] env[62753]: _type = "Task" [ 924.043646] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.045380] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1774591a-6afc-4f2b-b50f-3308a25f93dd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.055200] env[62753]: DEBUG oslo_concurrency.lockutils [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "628239e8-a447-4183-bed9-5550e53fda2b" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 924.055200] env[62753]: DEBUG oslo_concurrency.lockutils [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "628239e8-a447-4183-bed9-5550e53fda2b" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 924.063654] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332398, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.085915] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 924.086246] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 924.086480] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Deleting the datastore file [datastore2] 50e1c1a1-7b8f-49cd-932d-03c920209634 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 924.087285] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f67906c4-19bb-49ef-8052-4bf9d1c97e9e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.096444] env[62753]: DEBUG oslo_vmware.api [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for the task: (returnval){ [ 924.096444] env[62753]: value = "task-1332400" [ 924.096444] env[62753]: _type = "Task" [ 924.096444] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.107406] env[62753]: DEBUG oslo_vmware.api [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332400, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.460303] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332395, 'name': ReconfigVM_Task, 'duration_secs': 0.60069} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.460634] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Reconfigured VM instance instance-0000004c to attach disk [datastore1] a401909c-5978-492b-a4fa-b17d38a5af44/a401909c-5978-492b-a4fa-b17d38a5af44.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.461190] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2e37bba9-ffec-43a3-9131-dd76cc27e5f7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.465061] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1da73555-fc82-471a-8a37-42117658164c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.470517] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 924.470517] env[62753]: value = "task-1332401" [ 924.470517] env[62753]: _type = "Task" [ 924.470517] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.476400] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-243c99f8-f4f6-4f56-8a46-06dbe5229987 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.512254] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332401, 'name': Rename_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.513702] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf02e80e-6ecf-4b5f-b882-96db96f083f2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.522372] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d40cfac5-464d-41e8-bdb8-0daebed479c4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.539569] env[62753]: DEBUG nova.compute.provider_tree [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 924.557919] env[62753]: DEBUG nova.compute.manager [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 924.560867] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332398, 'name': ReconfigVM_Task, 'duration_secs': 0.323984} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.561499] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Reconfigured VM instance instance-0000004d to attach disk [datastore1] b45bfcb4-b2cd-468c-9595-62c155d520fb/b45bfcb4-b2cd-468c-9595-62c155d520fb.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 924.562418] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2f24243e-fec2-482c-838c-40ebeaff4ebb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.572058] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 924.572058] env[62753]: value = "task-1332402" [ 924.572058] env[62753]: _type = "Task" [ 924.572058] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 924.581875] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332402, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 924.606840] env[62753]: DEBUG oslo_vmware.api [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Task: {'id': task-1332400, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.150342} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.607130] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 924.607365] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 924.607504] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 924.607800] env[62753]: INFO nova.compute.manager [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Took 1.14 seconds to destroy the instance on the hypervisor. [ 924.607909] env[62753]: DEBUG oslo.service.loopingcall [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 924.608127] env[62753]: DEBUG nova.compute.manager [-] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 924.608226] env[62753]: DEBUG nova.network.neutron [-] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 924.626700] env[62753]: DEBUG nova.network.neutron [-] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 924.891719] env[62753]: DEBUG nova.compute.manager [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 924.895921] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-839c2eaf-a153-488b-b789-289157e4b1f2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.988171] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332401, 'name': Rename_Task, 'duration_secs': 0.220584} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 924.988171] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 924.988171] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-7570a2eb-b654-4083-a8d5-d11b582108e9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 924.996340] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 924.996340] env[62753]: value = "task-1332403" [ 924.996340] env[62753]: _type = "Task" [ 924.996340] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.006176] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332403, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.043357] env[62753]: DEBUG nova.scheduler.client.report [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 925.084824] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332402, 'name': Rename_Task, 'duration_secs': 0.171684} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.085950] env[62753]: DEBUG oslo_concurrency.lockutils [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.086265] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 925.086540] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8963b193-5697-4d38-91d7-94df4965af03 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.096232] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 925.096232] env[62753]: value = "task-1332404" [ 925.096232] env[62753]: _type = "Task" [ 925.096232] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.106223] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332404, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.129437] env[62753]: DEBUG nova.network.neutron [-] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 925.404999] env[62753]: INFO nova.compute.manager [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] instance snapshotting [ 925.408095] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e29442fc-95ef-46a1-9612-618367f92de7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.430409] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb1b15e2-591e-456d-88dd-3b7e9f056959 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.508416] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332403, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.548347] env[62753]: DEBUG oslo_concurrency.lockutils [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.334s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.548912] env[62753]: DEBUG nova.compute.manager [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 925.551657] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.262s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.553099] env[62753]: INFO nova.compute.claims [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 925.610099] env[62753]: DEBUG oslo_vmware.api [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332404, 'name': PowerOnVM_Task, 'duration_secs': 0.508997} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 925.610530] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 925.610855] env[62753]: INFO nova.compute.manager [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Took 7.75 seconds to spawn the instance on the hypervisor. [ 925.611197] env[62753]: DEBUG nova.compute.manager [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 925.612720] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60933cbb-0fe3-4038-81c5-e946830aa8a9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.632587] env[62753]: INFO nova.compute.manager [-] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Took 1.02 seconds to deallocate network for instance. [ 925.778037] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Acquiring lock "c5dae9a3-052b-4f4c-ac79-84aeaded457a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.778277] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Lock "c5dae9a3-052b-4f4c-ac79-84aeaded457a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.779242] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Acquiring lock "c5dae9a3-052b-4f4c-ac79-84aeaded457a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 925.779242] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Lock "c5dae9a3-052b-4f4c-ac79-84aeaded457a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 925.779242] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Lock "c5dae9a3-052b-4f4c-ac79-84aeaded457a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 925.781379] env[62753]: INFO nova.compute.manager [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Terminating instance [ 925.783595] env[62753]: DEBUG nova.compute.manager [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 925.783809] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 925.784063] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-293f2217-cf49-45d9-b2df-87d46dfea1b9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.793632] env[62753]: DEBUG oslo_vmware.api [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Waiting for the task: (returnval){ [ 925.793632] env[62753]: value = "task-1332405" [ 925.793632] env[62753]: _type = "Task" [ 925.793632] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.802818] env[62753]: DEBUG oslo_vmware.api [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332405, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 925.941940] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Creating Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 925.942306] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-f7896553-7c1e-418e-978c-c2ab31b66e14 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 925.950893] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 925.950893] env[62753]: value = "task-1332406" [ 925.950893] env[62753]: _type = "Task" [ 925.950893] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 925.960856] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332406, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.009960] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332403, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.059227] env[62753]: DEBUG nova.compute.utils [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 926.063216] env[62753]: DEBUG nova.compute.manager [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 926.063442] env[62753]: DEBUG nova.network.neutron [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 926.117249] env[62753]: DEBUG nova.policy [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '4603d0d0a2a14e419f6c34d770e562e5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '4b01e80a17e843a98ee662157230ae17', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 926.132466] env[62753]: INFO nova.compute.manager [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Took 34.07 seconds to build instance. [ 926.140113] env[62753]: DEBUG oslo_concurrency.lockutils [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 926.304910] env[62753]: DEBUG oslo_vmware.api [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332405, 'name': PowerOffVM_Task, 'duration_secs': 0.232462} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.305193] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 926.305404] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Volume detach. Driver type: vmdk {{(pid=62753) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 926.305595] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284559', 'volume_id': '02056526-1ec1-42b1-96d8-e237f65687ce', 'name': 'volume-02056526-1ec1-42b1-96d8-e237f65687ce', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c5dae9a3-052b-4f4c-ac79-84aeaded457a', 'attached_at': '', 'detached_at': '', 'volume_id': '02056526-1ec1-42b1-96d8-e237f65687ce', 'serial': '02056526-1ec1-42b1-96d8-e237f65687ce'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 926.306930] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-823ccf53-4494-4459-b6c4-e97f742be627 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.328482] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1579dfdc-ed6d-4598-afd3-e780e2823ca8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.336589] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24e7a465-5f44-4b49-bace-1a2791be8e9a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.355071] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a88f6061-1333-4a16-80ee-7fd8c253bd72 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.371174] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] The volume has not been displaced from its original location: [datastore2] volume-02056526-1ec1-42b1-96d8-e237f65687ce/volume-02056526-1ec1-42b1-96d8-e237f65687ce.vmdk. No consolidation needed. {{(pid=62753) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 926.376541] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Reconfiguring VM instance instance-00000040 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 926.376889] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a11792d8-bc09-43b8-a640-9812a5906018 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.390523] env[62753]: DEBUG nova.network.neutron [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Successfully created port: 167ccdaa-b799-422b-b031-ea36de9e571a {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 926.396071] env[62753]: DEBUG oslo_vmware.api [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Waiting for the task: (returnval){ [ 926.396071] env[62753]: value = "task-1332407" [ 926.396071] env[62753]: _type = "Task" [ 926.396071] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.406542] env[62753]: DEBUG oslo_vmware.api [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332407, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.461709] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332406, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.510829] env[62753]: DEBUG oslo_vmware.api [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332403, 'name': PowerOnVM_Task, 'duration_secs': 1.188097} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.511256] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 926.511574] env[62753]: INFO nova.compute.manager [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Took 11.04 seconds to spawn the instance on the hypervisor. [ 926.511961] env[62753]: DEBUG nova.compute.manager [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 926.513080] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc774d26-a555-4961-9275-c8935fea63e6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.564305] env[62753]: DEBUG nova.compute.manager [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 926.634544] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd56d291-9197-4983-95f5-40d501099bb8 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "b45bfcb4-b2cd-468c-9595-62c155d520fb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 35.628s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 926.658822] env[62753]: DEBUG nova.network.neutron [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Successfully created port: 6ea2368a-1e67-4403-9a24-10736cc87080 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 926.761374] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8f5f8f6-1396-42a3-8f9c-f50f6ca09ba8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.770355] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef338e61-e6df-4c72-b408-42ac67c40644 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.810134] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd9eee7c-4a9f-4af6-906c-8ed5bbb1d87b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.821499] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be03c6bd-f127-45ba-83d0-c8bad76c0981 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.838376] env[62753]: DEBUG nova.compute.provider_tree [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 926.906983] env[62753]: DEBUG oslo_vmware.api [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332407, 'name': ReconfigVM_Task, 'duration_secs': 0.283753} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 926.907337] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Reconfigured VM instance instance-00000040 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 926.912319] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6946f6a5-2e01-4e47-9531-6b3ff44306b8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 926.929366] env[62753]: DEBUG oslo_vmware.api [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Waiting for the task: (returnval){ [ 926.929366] env[62753]: value = "task-1332408" [ 926.929366] env[62753]: _type = "Task" [ 926.929366] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 926.940963] env[62753]: DEBUG oslo_vmware.api [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332408, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 926.962257] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332406, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.037051] env[62753]: INFO nova.compute.manager [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Took 36.46 seconds to build instance. [ 927.279324] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "b45bfcb4-b2cd-468c-9595-62c155d520fb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.279324] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "b45bfcb4-b2cd-468c-9595-62c155d520fb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.279324] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "b45bfcb4-b2cd-468c-9595-62c155d520fb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 927.279591] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "b45bfcb4-b2cd-468c-9595-62c155d520fb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.279647] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "b45bfcb4-b2cd-468c-9595-62c155d520fb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.281920] env[62753]: INFO nova.compute.manager [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Terminating instance [ 927.284166] env[62753]: DEBUG nova.compute.manager [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 927.284378] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 927.285256] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-333b7b5b-803f-4034-a0ee-53fe177ec655 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.293872] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 927.294144] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e6b77095-82f3-4986-87a2-15bf2d48e42e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.301255] env[62753]: DEBUG oslo_vmware.api [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 927.301255] env[62753]: value = "task-1332409" [ 927.301255] env[62753]: _type = "Task" [ 927.301255] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.309517] env[62753]: DEBUG oslo_vmware.api [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332409, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.341408] env[62753]: DEBUG nova.scheduler.client.report [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 927.439994] env[62753]: DEBUG oslo_vmware.api [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332408, 'name': ReconfigVM_Task, 'duration_secs': 0.177865} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.440341] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284559', 'volume_id': '02056526-1ec1-42b1-96d8-e237f65687ce', 'name': 'volume-02056526-1ec1-42b1-96d8-e237f65687ce', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'c5dae9a3-052b-4f4c-ac79-84aeaded457a', 'attached_at': '', 'detached_at': '', 'volume_id': '02056526-1ec1-42b1-96d8-e237f65687ce', 'serial': '02056526-1ec1-42b1-96d8-e237f65687ce'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 927.440649] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 927.441515] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a5debba-9782-4fd8-aa60-12da2530b04e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.448677] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 927.448924] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e5f41d5d-e2c9-48bd-b221-46f8b6d8570c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.460627] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332406, 'name': CreateSnapshot_Task, 'duration_secs': 1.258229} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 927.460876] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Created Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 927.461664] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8df4872-a3ce-4980-ac6e-2d08e14ef82d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.534040] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 927.534040] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 927.534281] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Deleting the datastore file [datastore2] c5dae9a3-052b-4f4c-ac79-84aeaded457a {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 927.534520] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-29be61ee-e58d-439c-85a6-a70f1ae8be44 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.538149] env[62753]: DEBUG oslo_concurrency.lockutils [None req-436b3a1b-0aa3-4a7f-9c11-05707ad51ae3 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "a401909c-5978-492b-a4fa-b17d38a5af44" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 43.671s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.541715] env[62753]: DEBUG oslo_vmware.api [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Waiting for the task: (returnval){ [ 927.541715] env[62753]: value = "task-1332411" [ 927.541715] env[62753]: _type = "Task" [ 927.541715] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 927.552031] env[62753]: DEBUG oslo_vmware.api [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332411, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.575729] env[62753]: DEBUG nova.compute.manager [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 927.597369] env[62753]: DEBUG nova.virt.hardware [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 927.597649] env[62753]: DEBUG nova.virt.hardware [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 927.597829] env[62753]: DEBUG nova.virt.hardware [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 927.598035] env[62753]: DEBUG nova.virt.hardware [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 927.598206] env[62753]: DEBUG nova.virt.hardware [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 927.598375] env[62753]: DEBUG nova.virt.hardware [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 927.598597] env[62753]: DEBUG nova.virt.hardware [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 927.598777] env[62753]: DEBUG nova.virt.hardware [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 927.598965] env[62753]: DEBUG nova.virt.hardware [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 927.599171] env[62753]: DEBUG nova.virt.hardware [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 927.599401] env[62753]: DEBUG nova.virt.hardware [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 927.600377] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cf33f8d-6d64-4b37-98a2-1d7d84ebb1b1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.609530] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e68a1e8d-61c0-4eea-a153-567dc1bc0591 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.811920] env[62753]: DEBUG oslo_vmware.api [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332409, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 927.845790] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.294s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 927.846391] env[62753]: DEBUG nova.compute.manager [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 927.849212] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.064s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 927.849586] env[62753]: DEBUG nova.objects.instance [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lazy-loading 'resources' on Instance uuid 3bc05fdb-81be-4764-b9d9-01acd125b020 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 927.979431] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Creating linked-clone VM from snapshot {{(pid=62753) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 927.979840] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d030f243-4d29-4551-9864-6e91ec248d15 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 927.991589] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 927.991589] env[62753]: value = "task-1332412" [ 927.991589] env[62753]: _type = "Task" [ 927.991589] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.001132] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332412, 'name': CloneVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.056532] env[62753]: DEBUG oslo_vmware.api [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Task: {'id': task-1332411, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.089719} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.056532] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 928.058627] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 928.058627] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 928.058627] env[62753]: INFO nova.compute.manager [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Took 2.27 seconds to destroy the instance on the hypervisor. [ 928.058627] env[62753]: DEBUG oslo.service.loopingcall [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 928.058627] env[62753]: DEBUG nova.compute.manager [-] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 928.058627] env[62753]: DEBUG nova.network.neutron [-] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 928.116724] env[62753]: DEBUG nova.compute.manager [req-757e1275-52f3-4a5a-8c74-cbd3f0d9b397 req-0a850eab-a43b-4083-9b0b-a073bb93ba88 service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Received event network-vif-plugged-167ccdaa-b799-422b-b031-ea36de9e571a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.116944] env[62753]: DEBUG oslo_concurrency.lockutils [req-757e1275-52f3-4a5a-8c74-cbd3f0d9b397 req-0a850eab-a43b-4083-9b0b-a073bb93ba88 service nova] Acquiring lock "f8470f38-0150-41c6-a3ee-aabb5e68d98b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 928.117490] env[62753]: DEBUG oslo_concurrency.lockutils [req-757e1275-52f3-4a5a-8c74-cbd3f0d9b397 req-0a850eab-a43b-4083-9b0b-a073bb93ba88 service nova] Lock "f8470f38-0150-41c6-a3ee-aabb5e68d98b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 928.117634] env[62753]: DEBUG oslo_concurrency.lockutils [req-757e1275-52f3-4a5a-8c74-cbd3f0d9b397 req-0a850eab-a43b-4083-9b0b-a073bb93ba88 service nova] Lock "f8470f38-0150-41c6-a3ee-aabb5e68d98b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 928.117815] env[62753]: DEBUG nova.compute.manager [req-757e1275-52f3-4a5a-8c74-cbd3f0d9b397 req-0a850eab-a43b-4083-9b0b-a073bb93ba88 service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] No waiting events found dispatching network-vif-plugged-167ccdaa-b799-422b-b031-ea36de9e571a {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 928.117991] env[62753]: WARNING nova.compute.manager [req-757e1275-52f3-4a5a-8c74-cbd3f0d9b397 req-0a850eab-a43b-4083-9b0b-a073bb93ba88 service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Received unexpected event network-vif-plugged-167ccdaa-b799-422b-b031-ea36de9e571a for instance with vm_state building and task_state spawning. [ 928.190421] env[62753]: DEBUG nova.network.neutron [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Successfully updated port: 167ccdaa-b799-422b-b031-ea36de9e571a {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 928.316242] env[62753]: DEBUG oslo_vmware.api [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332409, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.352783] env[62753]: DEBUG nova.compute.utils [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 928.358237] env[62753]: DEBUG nova.compute.manager [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 928.358639] env[62753]: DEBUG nova.network.neutron [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 928.413616] env[62753]: DEBUG nova.policy [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4dbc63095b0464fa0d2de86e3cf170e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5636da907ab343f9b42dbbd903d32283', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 928.509641] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332412, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.515869] env[62753]: DEBUG oslo_vmware.rw_handles [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520cfbbc-db58-305e-9b68-0096aff0aa7d/disk-0.vmdk. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 928.516790] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05e5b1a7-55dd-43fd-bd31-d7b79971959f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.527337] env[62753]: DEBUG oslo_vmware.rw_handles [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520cfbbc-db58-305e-9b68-0096aff0aa7d/disk-0.vmdk is in state: ready. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 928.527543] env[62753]: ERROR oslo_vmware.rw_handles [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520cfbbc-db58-305e-9b68-0096aff0aa7d/disk-0.vmdk due to incomplete transfer. [ 928.527800] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-4b1cdcbc-3493-41ea-b5c9-3996b8f0f32b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.538183] env[62753]: DEBUG oslo_vmware.rw_handles [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/520cfbbc-db58-305e-9b68-0096aff0aa7d/disk-0.vmdk. {{(pid=62753) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 928.539861] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Uploaded image 061f0060-350a-4bf7-a725-58acba59de2b to the Glance image server {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 928.541470] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Destroying the VM {{(pid=62753) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 928.544388] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-37857b4c-aa64-427d-be7d-34c6b2985d01 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.553048] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 928.553048] env[62753]: value = "task-1332413" [ 928.553048] env[62753]: _type = "Task" [ 928.553048] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.564687] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332413, 'name': Destroy_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.623899] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fbaa0e9-453d-45b4-b173-f4e047838a2c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.633505] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35e54c73-b235-4733-92dc-409b86eb6f99 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.670404] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a7ee21b-73e3-4bbd-9023-98431de58512 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.679270] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4eb5ec02-35e2-4f36-9180-730a796a9f36 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.696814] env[62753]: DEBUG nova.compute.provider_tree [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 928.705881] env[62753]: DEBUG nova.network.neutron [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Successfully created port: f433dba2-2032-4957-8ba6-06568d931ccc {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 928.818970] env[62753]: DEBUG oslo_vmware.api [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332409, 'name': PowerOffVM_Task, 'duration_secs': 1.100493} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 928.820472] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 928.820472] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 928.820472] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a9eca689-b161-41ad-b6de-44a75b4c32b3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.860697] env[62753]: DEBUG nova.compute.manager [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 928.898797] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 928.898797] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 928.898797] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Deleting the datastore file [datastore1] b45bfcb4-b2cd-468c-9595-62c155d520fb {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 928.898797] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-30bdc2a6-ba9f-4643-840c-c02a206a2825 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 928.905257] env[62753]: DEBUG oslo_vmware.api [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 928.905257] env[62753]: value = "task-1332415" [ 928.905257] env[62753]: _type = "Task" [ 928.905257] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 928.914655] env[62753]: DEBUG oslo_vmware.api [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332415, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 928.982876] env[62753]: DEBUG nova.compute.manager [req-e47623ec-54b7-4669-9b96-dc2df33eb6fc req-fd12568b-56de-4161-b91d-54a5a52da53f service nova] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Received event network-vif-deleted-e44dcb1f-60ee-409a-805b-0c795852f967 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 928.983187] env[62753]: INFO nova.compute.manager [req-e47623ec-54b7-4669-9b96-dc2df33eb6fc req-fd12568b-56de-4161-b91d-54a5a52da53f service nova] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Neutron deleted interface e44dcb1f-60ee-409a-805b-0c795852f967; detaching it from the instance and deleting it from the info cache [ 928.983341] env[62753]: DEBUG nova.network.neutron [req-e47623ec-54b7-4669-9b96-dc2df33eb6fc req-fd12568b-56de-4161-b91d-54a5a52da53f service nova] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.006319] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332412, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.066176] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332413, 'name': Destroy_Task, 'duration_secs': 0.504192} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.066486] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Destroyed the VM [ 929.069113] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Deleting Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 929.069113] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-5d5de052-e450-4f8d-aa24-62f66b163871 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.078430] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 929.078430] env[62753]: value = "task-1332416" [ 929.078430] env[62753]: _type = "Task" [ 929.078430] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 929.088961] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332416, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.199776] env[62753]: DEBUG nova.scheduler.client.report [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 929.417307] env[62753]: DEBUG oslo_vmware.api [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332415, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.141877} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 929.417602] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 929.417792] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 929.417978] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 929.418212] env[62753]: INFO nova.compute.manager [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Took 2.13 seconds to destroy the instance on the hypervisor. [ 929.418567] env[62753]: DEBUG oslo.service.loopingcall [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 929.418841] env[62753]: DEBUG nova.compute.manager [-] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 929.418841] env[62753]: DEBUG nova.network.neutron [-] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 929.451436] env[62753]: DEBUG nova.network.neutron [-] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 929.488927] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d98aece9-c4e8-4874-9aa3-2409bbb2cd1a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.501959] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7034620-fe38-4010-aaf5-2906f96279e6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.526979] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332412, 'name': CloneVM_Task} progress is 95%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.535914] env[62753]: DEBUG nova.compute.manager [req-e47623ec-54b7-4669-9b96-dc2df33eb6fc req-fd12568b-56de-4161-b91d-54a5a52da53f service nova] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Detach interface failed, port_id=e44dcb1f-60ee-409a-805b-0c795852f967, reason: Instance c5dae9a3-052b-4f4c-ac79-84aeaded457a could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 929.588567] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332416, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 929.705700] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.856s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 929.708074] env[62753]: DEBUG oslo_concurrency.lockutils [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.622s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 929.709602] env[62753]: INFO nova.compute.claims [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 929.726912] env[62753]: INFO nova.scheduler.client.report [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Deleted allocations for instance 3bc05fdb-81be-4764-b9d9-01acd125b020 [ 929.801874] env[62753]: DEBUG nova.compute.manager [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Stashing vm_state: active {{(pid=62753) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 929.870470] env[62753]: DEBUG nova.compute.manager [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 929.895530] env[62753]: DEBUG nova.virt.hardware [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 929.895835] env[62753]: DEBUG nova.virt.hardware [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 929.896050] env[62753]: DEBUG nova.virt.hardware [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 929.896307] env[62753]: DEBUG nova.virt.hardware [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 929.896495] env[62753]: DEBUG nova.virt.hardware [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 929.896682] env[62753]: DEBUG nova.virt.hardware [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 929.896940] env[62753]: DEBUG nova.virt.hardware [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 929.897184] env[62753]: DEBUG nova.virt.hardware [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 929.897406] env[62753]: DEBUG nova.virt.hardware [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 929.897601] env[62753]: DEBUG nova.virt.hardware [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 929.897782] env[62753]: DEBUG nova.virt.hardware [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 929.898944] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-632589f3-9e12-48be-9ad7-f079bccfbff1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.908990] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e5709f09-2d1c-41a3-8a4c-fe8975ec4f8c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 929.954045] env[62753]: INFO nova.compute.manager [-] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Took 1.90 seconds to deallocate network for instance. [ 930.007554] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332412, 'name': CloneVM_Task, 'duration_secs': 1.82736} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.007863] env[62753]: INFO nova.virt.vmwareapi.vmops [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Created linked-clone VM from snapshot [ 930.008671] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7712a435-e2b3-4e37-9ebe-6d0ac20d9f7e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.017787] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Uploading image 673e56af-bdea-458e-b52c-3abfd1149b15 {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 930.042689] env[62753]: DEBUG oslo_vmware.rw_handles [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 930.042689] env[62753]: value = "vm-284625" [ 930.042689] env[62753]: _type = "VirtualMachine" [ 930.042689] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 930.042982] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-20a0eb94-4d71-4fb0-a708-05932cd2a7ba {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.052031] env[62753]: DEBUG oslo_vmware.rw_handles [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lease: (returnval){ [ 930.052031] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529cfad8-e3cb-f193-9e8c-bbccedb2975d" [ 930.052031] env[62753]: _type = "HttpNfcLease" [ 930.052031] env[62753]: } obtained for exporting VM: (result){ [ 930.052031] env[62753]: value = "vm-284625" [ 930.052031] env[62753]: _type = "VirtualMachine" [ 930.052031] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 930.052031] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the lease: (returnval){ [ 930.052031] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529cfad8-e3cb-f193-9e8c-bbccedb2975d" [ 930.052031] env[62753]: _type = "HttpNfcLease" [ 930.052031] env[62753]: } to be ready. {{(pid=62753) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 930.059533] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 930.059533] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529cfad8-e3cb-f193-9e8c-bbccedb2975d" [ 930.059533] env[62753]: _type = "HttpNfcLease" [ 930.059533] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 930.087935] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332416, 'name': RemoveSnapshot_Task, 'duration_secs': 0.863557} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 930.088244] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Deleted Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 930.088530] env[62753]: DEBUG nova.compute.manager [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 930.089328] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75974e14-08be-423a-a84f-e823d670b7ea {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.145637] env[62753]: DEBUG nova.compute.manager [req-216914b8-23e0-4429-bc97-c424e65233cf req-0abb27f9-21bb-4e3b-a163-9158e2c6e10e service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Received event network-changed-167ccdaa-b799-422b-b031-ea36de9e571a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 930.145897] env[62753]: DEBUG nova.compute.manager [req-216914b8-23e0-4429-bc97-c424e65233cf req-0abb27f9-21bb-4e3b-a163-9158e2c6e10e service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Refreshing instance network info cache due to event network-changed-167ccdaa-b799-422b-b031-ea36de9e571a. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 930.146148] env[62753]: DEBUG oslo_concurrency.lockutils [req-216914b8-23e0-4429-bc97-c424e65233cf req-0abb27f9-21bb-4e3b-a163-9158e2c6e10e service nova] Acquiring lock "refresh_cache-f8470f38-0150-41c6-a3ee-aabb5e68d98b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.146698] env[62753]: DEBUG oslo_concurrency.lockutils [req-216914b8-23e0-4429-bc97-c424e65233cf req-0abb27f9-21bb-4e3b-a163-9158e2c6e10e service nova] Acquired lock "refresh_cache-f8470f38-0150-41c6-a3ee-aabb5e68d98b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.146698] env[62753]: DEBUG nova.network.neutron [req-216914b8-23e0-4429-bc97-c424e65233cf req-0abb27f9-21bb-4e3b-a163-9158e2c6e10e service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Refreshing network info cache for port 167ccdaa-b799-422b-b031-ea36de9e571a {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 930.235974] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b26b4851-83fa-4f6a-9980-278a4f4fa691 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "3bc05fdb-81be-4764-b9d9-01acd125b020" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.016s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 930.237245] env[62753]: DEBUG oslo_concurrency.lockutils [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] Acquired lock "3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.238295] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0cccc58c-67b3-4ae6-b020-be67c9cb171e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.248964] env[62753]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 930.248964] env[62753]: DEBUG oslo_vmware.api [-] Fault list: [ManagedObjectNotFound] {{(pid=62753) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 930.249959] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e3864911-b788-49d5-9241-22ac8260a50d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.260742] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af8ff8e6-0e59-4133-81f9-315b01e73da5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.294708] env[62753]: ERROR root [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] Original exception being dropped: ['Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 377, in request_handler\n response = request(managed_object, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 586, in __call__\n return client.invoke(args, kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 728, in invoke\n result = self.send(soapenv, timeout=timeout)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 777, in send\n return self.process_reply(reply.message, None, None)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/suds/client.py", line 840, in process_reply\n raise WebFault(fault, replyroot)\n', "suds.WebFault: Server raised fault: 'The object 'vim.VirtualMachine:vm-284563' has already been deleted or has not been completely created'\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 301, in _invoke_api\n return api_method(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 480, in get_object_property\n props = get_object_properties(vim, moref, [property_name],\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/vim_util.py", line 360, in get_object_properties\n retrieve_result = vim.RetrievePropertiesEx(\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py", line 413, in request_handler\n raise exceptions.VimFaultException(fault_list, fault_string,\n', "oslo_vmware.exceptions.VimFaultException: The object 'vim.VirtualMachine:vm-284563' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-284563' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-284563'}\n", '\nDuring handling of the above exception, another exception occurred:\n\n', 'Traceback (most recent call last):\n', ' File "/opt/stack/nova/nova/virt/vmwareapi/session.py", line 123, in _call_method\n return self.invoke_api(module, method, self.vim, *args,\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 358, in invoke_api\n return _invoke_api(module, method, *args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 122, in func\n return evt.wait()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/event.py", line 124, in wait\n result = hub.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/eventlet/hubs/hub.py", line 310, in switch\n return self.greenlet.switch()\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/common/loopingcall.py", line 122, in _inner\n idle = self.f(*self.args, **self.kw)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 96, in _func\n result = f(*args, **kwargs)\n', ' File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py", line 341, in _invoke_api\n raise clazz(str(excep),\n', "oslo_vmware.exceptions.ManagedObjectNotFoundException: The object 'vim.VirtualMachine:vm-284563' has already been deleted or has not been completely created\nCause: Server raised fault: 'The object 'vim.VirtualMachine:vm-284563' has already been deleted or has not been completely created'\nFaults: [ManagedObjectNotFound]\nDetails: {'obj': 'vm-284563'}\n"]: nova.exception.InstanceNotFound: Instance 3bc05fdb-81be-4764-b9d9-01acd125b020 could not be found. [ 930.294976] env[62753]: DEBUG oslo_concurrency.lockutils [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] Releasing lock "3bc05fdb-81be-4764-b9d9-01acd125b020" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 930.295255] env[62753]: DEBUG nova.compute.manager [req-a0166516-e7aa-4c46-8f74-529bfebde47b req-60aa6842-8434-4f1b-b10a-e0853ae5fa32 service nova] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Detach interface failed, port_id=9713beba-767d-408c-bd7d-bef58dc3c405, reason: Instance 3bc05fdb-81be-4764-b9d9-01acd125b020 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 930.322026] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 930.332519] env[62753]: DEBUG nova.network.neutron [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Successfully updated port: f433dba2-2032-4957-8ba6-06568d931ccc {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 930.339886] env[62753]: DEBUG nova.network.neutron [-] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.434482] env[62753]: DEBUG nova.network.neutron [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Successfully updated port: 6ea2368a-1e67-4403-9a24-10736cc87080 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 930.497620] env[62753]: INFO nova.compute.manager [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Took 0.54 seconds to detach 1 volumes for instance. [ 930.500037] env[62753]: DEBUG nova.compute.manager [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Deleting volume: 02056526-1ec1-42b1-96d8-e237f65687ce {{(pid=62753) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 930.562398] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 930.562398] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529cfad8-e3cb-f193-9e8c-bbccedb2975d" [ 930.562398] env[62753]: _type = "HttpNfcLease" [ 930.562398] env[62753]: } is ready. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 930.562767] env[62753]: DEBUG oslo_vmware.rw_handles [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 930.562767] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529cfad8-e3cb-f193-9e8c-bbccedb2975d" [ 930.562767] env[62753]: _type = "HttpNfcLease" [ 930.562767] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 930.564433] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ef3fe61-9c5c-4f7a-8726-402556705385 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.573137] env[62753]: DEBUG oslo_vmware.rw_handles [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52431f05-8fa3-faad-6709-e9d66ebe5e64/disk-0.vmdk from lease info. {{(pid=62753) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 930.573530] env[62753]: DEBUG oslo_vmware.rw_handles [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52431f05-8fa3-faad-6709-e9d66ebe5e64/disk-0.vmdk for reading. {{(pid=62753) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 930.636500] env[62753]: INFO nova.compute.manager [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Shelve offloading [ 930.638160] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 930.640033] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-769b5bc9-9e76-4b34-8685-f26a327eaa03 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.653044] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 930.653044] env[62753]: value = "task-1332419" [ 930.653044] env[62753]: _type = "Task" [ 930.653044] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 930.662725] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] VM already powered off {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 930.662935] env[62753]: DEBUG nova.compute.manager [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 930.665360] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5292b8d0-bed0-4e54-8093-81c856566e43 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.670412] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.670588] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquired lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.670769] env[62753]: DEBUG nova.network.neutron [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 930.685222] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-25fe8813-59a5-4108-bd29-3c54e0e36f7b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.688181] env[62753]: DEBUG nova.network.neutron [req-216914b8-23e0-4429-bc97-c424e65233cf req-0abb27f9-21bb-4e3b-a163-9158e2c6e10e service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 930.836994] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "refresh_cache-8066aeb2-7be4-46e3-aba6-124fef342a49" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.837174] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "refresh_cache-8066aeb2-7be4-46e3-aba6-124fef342a49" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 930.837330] env[62753]: DEBUG nova.network.neutron [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 930.842545] env[62753]: INFO nova.compute.manager [-] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Took 1.42 seconds to deallocate network for instance. [ 930.857976] env[62753]: DEBUG nova.network.neutron [req-216914b8-23e0-4429-bc97-c424e65233cf req-0abb27f9-21bb-4e3b-a163-9158e2c6e10e service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 930.938618] env[62753]: DEBUG oslo_concurrency.lockutils [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquiring lock "refresh_cache-f8470f38-0150-41c6-a3ee-aabb5e68d98b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 930.974196] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-15d22200-3d69-4b77-ab64-206f53dd337d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 930.983667] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb85d380-3916-4189-af9a-f70f63ddfcea {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.016636] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad448ee7-ade1-4e41-a496-7a924e2dd48f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.026108] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-775f8397-38cc-4177-8705-624316610f5e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 931.033007] env[62753]: DEBUG nova.compute.manager [req-89045458-3d50-40f7-b895-769e3c5f671d req-0fda8cb1-0427-4923-8194-a14441e151d3 service nova] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Received event network-vif-deleted-91add90d-924f-43de-bead-d4f33e2af4ad {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.033409] env[62753]: DEBUG nova.compute.manager [req-89045458-3d50-40f7-b895-769e3c5f671d req-0fda8cb1-0427-4923-8194-a14441e151d3 service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Received event network-vif-plugged-6ea2368a-1e67-4403-9a24-10736cc87080 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.033614] env[62753]: DEBUG oslo_concurrency.lockutils [req-89045458-3d50-40f7-b895-769e3c5f671d req-0fda8cb1-0427-4923-8194-a14441e151d3 service nova] Acquiring lock "f8470f38-0150-41c6-a3ee-aabb5e68d98b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.033820] env[62753]: DEBUG oslo_concurrency.lockutils [req-89045458-3d50-40f7-b895-769e3c5f671d req-0fda8cb1-0427-4923-8194-a14441e151d3 service nova] Lock "f8470f38-0150-41c6-a3ee-aabb5e68d98b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 931.033983] env[62753]: DEBUG oslo_concurrency.lockutils [req-89045458-3d50-40f7-b895-769e3c5f671d req-0fda8cb1-0427-4923-8194-a14441e151d3 service nova] Lock "f8470f38-0150-41c6-a3ee-aabb5e68d98b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 931.034181] env[62753]: DEBUG nova.compute.manager [req-89045458-3d50-40f7-b895-769e3c5f671d req-0fda8cb1-0427-4923-8194-a14441e151d3 service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] No waiting events found dispatching network-vif-plugged-6ea2368a-1e67-4403-9a24-10736cc87080 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 931.034482] env[62753]: WARNING nova.compute.manager [req-89045458-3d50-40f7-b895-769e3c5f671d req-0fda8cb1-0427-4923-8194-a14441e151d3 service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Received unexpected event network-vif-plugged-6ea2368a-1e67-4403-9a24-10736cc87080 for instance with vm_state building and task_state spawning. [ 931.034716] env[62753]: DEBUG nova.compute.manager [req-89045458-3d50-40f7-b895-769e3c5f671d req-0fda8cb1-0427-4923-8194-a14441e151d3 service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Received event network-changed-6ea2368a-1e67-4403-9a24-10736cc87080 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 931.034958] env[62753]: DEBUG nova.compute.manager [req-89045458-3d50-40f7-b895-769e3c5f671d req-0fda8cb1-0427-4923-8194-a14441e151d3 service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Refreshing instance network info cache due to event network-changed-6ea2368a-1e67-4403-9a24-10736cc87080. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 931.035202] env[62753]: DEBUG oslo_concurrency.lockutils [req-89045458-3d50-40f7-b895-769e3c5f671d req-0fda8cb1-0427-4923-8194-a14441e151d3 service nova] Acquiring lock "refresh_cache-f8470f38-0150-41c6-a3ee-aabb5e68d98b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 931.047229] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.047748] env[62753]: DEBUG nova.compute.provider_tree [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 931.352017] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 931.361900] env[62753]: DEBUG oslo_concurrency.lockutils [req-216914b8-23e0-4429-bc97-c424e65233cf req-0abb27f9-21bb-4e3b-a163-9158e2c6e10e service nova] Releasing lock "refresh_cache-f8470f38-0150-41c6-a3ee-aabb5e68d98b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 931.363716] env[62753]: DEBUG oslo_concurrency.lockutils [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquired lock "refresh_cache-f8470f38-0150-41c6-a3ee-aabb5e68d98b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 931.365757] env[62753]: DEBUG nova.network.neutron [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 931.388134] env[62753]: DEBUG nova.network.neutron [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 931.447597] env[62753]: DEBUG nova.network.neutron [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Updating instance_info_cache with network_info: [{"id": "7fb046f4-e470-4f6d-8fa3-73f3b0157020", "address": "fa:16:3e:6f:26:2c", "network": {"id": "fc646f7f-70ba-4671-8476-3b0ece1e9041", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-853174775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d0eb0b3ba504f59a85cb021da2b47b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fb046f4-e4", "ovs_interfaceid": "7fb046f4-e470-4f6d-8fa3-73f3b0157020", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.551203] env[62753]: DEBUG nova.scheduler.client.report [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 931.676218] env[62753]: DEBUG nova.network.neutron [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Updating instance_info_cache with network_info: [{"id": "f433dba2-2032-4957-8ba6-06568d931ccc", "address": "fa:16:3e:97:cc:94", "network": {"id": "35ed0ec0-821f-45b6-bc2d-015626bfc0a2", "bridge": "br-int", "label": "tempest-ServersTestJSON-444726172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5636da907ab343f9b42dbbd903d32283", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf433dba2-20", "ovs_interfaceid": "f433dba2-2032-4957-8ba6-06568d931ccc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 931.916385] env[62753]: DEBUG nova.network.neutron [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 931.951144] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Releasing lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.056463] env[62753]: DEBUG oslo_concurrency.lockutils [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.348s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.057042] env[62753]: DEBUG nova.compute.manager [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 932.059650] env[62753]: DEBUG oslo_concurrency.lockutils [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.920s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.059926] env[62753]: DEBUG nova.objects.instance [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lazy-loading 'resources' on Instance uuid 50e1c1a1-7b8f-49cd-932d-03c920209634 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 932.180928] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "refresh_cache-8066aeb2-7be4-46e3-aba6-124fef342a49" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.181432] env[62753]: DEBUG nova.compute.manager [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Instance network_info: |[{"id": "f433dba2-2032-4957-8ba6-06568d931ccc", "address": "fa:16:3e:97:cc:94", "network": {"id": "35ed0ec0-821f-45b6-bc2d-015626bfc0a2", "bridge": "br-int", "label": "tempest-ServersTestJSON-444726172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5636da907ab343f9b42dbbd903d32283", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf433dba2-20", "ovs_interfaceid": "f433dba2-2032-4957-8ba6-06568d931ccc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 932.181911] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:97:cc:94', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2624812a-9f9c-461d-8b5f-79bea90c7ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f433dba2-2032-4957-8ba6-06568d931ccc', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 932.189634] env[62753]: DEBUG oslo.service.loopingcall [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.190669] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 932.190912] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6bd2ac32-4de9-4494-9f97-220474387dca {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.213932] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 932.213932] env[62753]: value = "task-1332420" [ 932.213932] env[62753]: _type = "Task" [ 932.213932] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.222663] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332420, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.230465] env[62753]: DEBUG nova.network.neutron [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Updating instance_info_cache with network_info: [{"id": "167ccdaa-b799-422b-b031-ea36de9e571a", "address": "fa:16:3e:7f:2d:56", "network": {"id": "61486b9e-1554-4e34-990f-38a03833f61f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-855522933", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.39", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b01e80a17e843a98ee662157230ae17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b8137fc-f23d-49b1-b19c-3123a5588f34", "external-id": "nsx-vlan-transportzone-709", "segmentation_id": 709, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap167ccdaa-b7", "ovs_interfaceid": "167ccdaa-b799-422b-b031-ea36de9e571a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6ea2368a-1e67-4403-9a24-10736cc87080", "address": "fa:16:3e:24:f0:ea", "network": {"id": "c2c606e7-dc7b-4aa4-9abc-b5299f6b4660", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-455294558", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.127", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4b01e80a17e843a98ee662157230ae17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e839c46-1ae9-43b7-9518-8f18f48100dd", "external-id": "nsx-vlan-transportzone-666", "segmentation_id": 666, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ea2368a-1e", "ovs_interfaceid": "6ea2368a-1e67-4403-9a24-10736cc87080", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 932.566465] env[62753]: DEBUG nova.compute.utils [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 932.569056] env[62753]: DEBUG nova.compute.manager [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 932.569056] env[62753]: DEBUG nova.network.neutron [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 932.623169] env[62753]: DEBUG nova.policy [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae52823f7c93454e8089261c3ca44321', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d52d6eaee934be5ab0e0003df1ce316', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 932.728386] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332420, 'name': CreateVM_Task, 'duration_secs': 0.372609} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.728755] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 932.729597] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.729902] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.730333] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 932.730697] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-56f69a51-4494-4063-8450-b8f6e2714b8c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.736027] env[62753]: DEBUG oslo_concurrency.lockutils [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Releasing lock "refresh_cache-f8470f38-0150-41c6-a3ee-aabb5e68d98b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.736027] env[62753]: DEBUG nova.compute.manager [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Instance network_info: |[{"id": "167ccdaa-b799-422b-b031-ea36de9e571a", "address": "fa:16:3e:7f:2d:56", "network": {"id": "61486b9e-1554-4e34-990f-38a03833f61f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-855522933", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.39", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b01e80a17e843a98ee662157230ae17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b8137fc-f23d-49b1-b19c-3123a5588f34", "external-id": "nsx-vlan-transportzone-709", "segmentation_id": 709, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap167ccdaa-b7", "ovs_interfaceid": "167ccdaa-b799-422b-b031-ea36de9e571a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6ea2368a-1e67-4403-9a24-10736cc87080", "address": "fa:16:3e:24:f0:ea", "network": {"id": "c2c606e7-dc7b-4aa4-9abc-b5299f6b4660", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-455294558", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.127", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4b01e80a17e843a98ee662157230ae17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e839c46-1ae9-43b7-9518-8f18f48100dd", "external-id": "nsx-vlan-transportzone-666", "segmentation_id": 666, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ea2368a-1e", "ovs_interfaceid": "6ea2368a-1e67-4403-9a24-10736cc87080", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 932.736027] env[62753]: DEBUG oslo_concurrency.lockutils [req-89045458-3d50-40f7-b895-769e3c5f671d req-0fda8cb1-0427-4923-8194-a14441e151d3 service nova] Acquired lock "refresh_cache-f8470f38-0150-41c6-a3ee-aabb5e68d98b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.736027] env[62753]: DEBUG nova.network.neutron [req-89045458-3d50-40f7-b895-769e3c5f671d req-0fda8cb1-0427-4923-8194-a14441e151d3 service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Refreshing network info cache for port 6ea2368a-1e67-4403-9a24-10736cc87080 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 932.736027] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:7f:2d:56', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '6b8137fc-f23d-49b1-b19c-3123a5588f34', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '167ccdaa-b799-422b-b031-ea36de9e571a', 'vif_model': 'vmxnet3'}, {'network_name': 'br-int', 'mac_address': 'fa:16:3e:24:f0:ea', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '5e839c46-1ae9-43b7-9518-8f18f48100dd', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6ea2368a-1e67-4403-9a24-10736cc87080', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 932.745602] env[62753]: DEBUG oslo.service.loopingcall [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 932.749978] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 932.751937] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-8cc8ff52-fb2b-4a23-b4b4-ad355181835f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.772493] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 932.772493] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c133c9-c5c0-41d5-6eb9-9990893b8ba1" [ 932.772493] env[62753]: _type = "Task" [ 932.772493] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.783011] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 932.783011] env[62753]: value = "task-1332421" [ 932.783011] env[62753]: _type = "Task" [ 932.783011] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.788056] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c133c9-c5c0-41d5-6eb9-9990893b8ba1, 'name': SearchDatastore_Task, 'duration_secs': 0.01295} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 932.793953] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 932.794654] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 932.795163] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.795613] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.797932] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 932.797932] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-b331ce3a-488b-4637-8d99-4ee434009a4d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.811233] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332421, 'name': CreateVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.815186] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 932.815468] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 932.818522] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e52e94bf-85f2-40be-a969-b48944888270 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.825813] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 932.825813] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dc994f-44b2-528c-25b0-db9c56975c73" [ 932.825813] env[62753]: _type = "Task" [ 932.825813] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 932.838582] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dc994f-44b2-528c-25b0-db9c56975c73, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 932.841541] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee686c63-8308-4926-a782-64922ba64cc0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.852631] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5a2c95-684c-4aee-8ca3-24d1a23e9f74 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.888617] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47f1a70c-16a2-49d0-a388-4c2cf6ffee81 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.899909] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64600a7d-3d2f-4bbb-a67e-3563f80b4515 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.914891] env[62753]: DEBUG nova.compute.provider_tree [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 932.917154] env[62753]: DEBUG nova.network.neutron [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Successfully created port: 800db109-09e3-45a4-8753-20d9911c0677 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 932.929548] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 932.930902] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a80b57ba-dce7-4d81-846d-a53b8b81885e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.940799] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 932.941101] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-1a0208a3-4dc7-433b-bd78-80846ae7b218 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 932.946538] env[62753]: DEBUG nova.compute.manager [req-9a784af9-536f-4f8c-b17a-691dc33880e4 req-998f8703-10c2-4bf4-ad6f-27670c7d7fb1 service nova] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Received event network-vif-plugged-f433dba2-2032-4957-8ba6-06568d931ccc {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.946898] env[62753]: DEBUG oslo_concurrency.lockutils [req-9a784af9-536f-4f8c-b17a-691dc33880e4 req-998f8703-10c2-4bf4-ad6f-27670c7d7fb1 service nova] Acquiring lock "8066aeb2-7be4-46e3-aba6-124fef342a49-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 932.947310] env[62753]: DEBUG oslo_concurrency.lockutils [req-9a784af9-536f-4f8c-b17a-691dc33880e4 req-998f8703-10c2-4bf4-ad6f-27670c7d7fb1 service nova] Lock "8066aeb2-7be4-46e3-aba6-124fef342a49-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 932.947620] env[62753]: DEBUG oslo_concurrency.lockutils [req-9a784af9-536f-4f8c-b17a-691dc33880e4 req-998f8703-10c2-4bf4-ad6f-27670c7d7fb1 service nova] Lock "8066aeb2-7be4-46e3-aba6-124fef342a49-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 932.947929] env[62753]: DEBUG nova.compute.manager [req-9a784af9-536f-4f8c-b17a-691dc33880e4 req-998f8703-10c2-4bf4-ad6f-27670c7d7fb1 service nova] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] No waiting events found dispatching network-vif-plugged-f433dba2-2032-4957-8ba6-06568d931ccc {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 932.948257] env[62753]: WARNING nova.compute.manager [req-9a784af9-536f-4f8c-b17a-691dc33880e4 req-998f8703-10c2-4bf4-ad6f-27670c7d7fb1 service nova] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Received unexpected event network-vif-plugged-f433dba2-2032-4957-8ba6-06568d931ccc for instance with vm_state building and task_state spawning. [ 932.948558] env[62753]: DEBUG nova.compute.manager [req-9a784af9-536f-4f8c-b17a-691dc33880e4 req-998f8703-10c2-4bf4-ad6f-27670c7d7fb1 service nova] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Received event network-changed-f433dba2-2032-4957-8ba6-06568d931ccc {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 932.948853] env[62753]: DEBUG nova.compute.manager [req-9a784af9-536f-4f8c-b17a-691dc33880e4 req-998f8703-10c2-4bf4-ad6f-27670c7d7fb1 service nova] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Refreshing instance network info cache due to event network-changed-f433dba2-2032-4957-8ba6-06568d931ccc. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 932.949202] env[62753]: DEBUG oslo_concurrency.lockutils [req-9a784af9-536f-4f8c-b17a-691dc33880e4 req-998f8703-10c2-4bf4-ad6f-27670c7d7fb1 service nova] Acquiring lock "refresh_cache-8066aeb2-7be4-46e3-aba6-124fef342a49" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 932.949476] env[62753]: DEBUG oslo_concurrency.lockutils [req-9a784af9-536f-4f8c-b17a-691dc33880e4 req-998f8703-10c2-4bf4-ad6f-27670c7d7fb1 service nova] Acquired lock "refresh_cache-8066aeb2-7be4-46e3-aba6-124fef342a49" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 932.949778] env[62753]: DEBUG nova.network.neutron [req-9a784af9-536f-4f8c-b17a-691dc33880e4 req-998f8703-10c2-4bf4-ad6f-27670c7d7fb1 service nova] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Refreshing network info cache for port f433dba2-2032-4957-8ba6-06568d931ccc {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 933.010313] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 933.010548] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 933.011077] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Deleting the datastore file [datastore1] 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 933.011077] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-9c3ca4ef-c952-41a0-9506-ec136b1dcd85 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.021179] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 933.021179] env[62753]: value = "task-1332423" [ 933.021179] env[62753]: _type = "Task" [ 933.021179] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.030257] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332423, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.074774] env[62753]: DEBUG nova.compute.manager [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 933.111064] env[62753]: DEBUG nova.compute.manager [req-770f2557-6f96-4d48-8621-64bb597378f8 req-29142090-51e1-4984-8ff5-95597d1601b4 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Received event network-vif-unplugged-7fb046f4-e470-4f6d-8fa3-73f3b0157020 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.111377] env[62753]: DEBUG oslo_concurrency.lockutils [req-770f2557-6f96-4d48-8621-64bb597378f8 req-29142090-51e1-4984-8ff5-95597d1601b4 service nova] Acquiring lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.111698] env[62753]: DEBUG oslo_concurrency.lockutils [req-770f2557-6f96-4d48-8621-64bb597378f8 req-29142090-51e1-4984-8ff5-95597d1601b4 service nova] Lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.111950] env[62753]: DEBUG oslo_concurrency.lockutils [req-770f2557-6f96-4d48-8621-64bb597378f8 req-29142090-51e1-4984-8ff5-95597d1601b4 service nova] Lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.112191] env[62753]: DEBUG nova.compute.manager [req-770f2557-6f96-4d48-8621-64bb597378f8 req-29142090-51e1-4984-8ff5-95597d1601b4 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] No waiting events found dispatching network-vif-unplugged-7fb046f4-e470-4f6d-8fa3-73f3b0157020 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 933.112413] env[62753]: WARNING nova.compute.manager [req-770f2557-6f96-4d48-8621-64bb597378f8 req-29142090-51e1-4984-8ff5-95597d1601b4 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Received unexpected event network-vif-unplugged-7fb046f4-e470-4f6d-8fa3-73f3b0157020 for instance with vm_state shelved and task_state shelving_offloading. [ 933.112623] env[62753]: DEBUG nova.compute.manager [req-770f2557-6f96-4d48-8621-64bb597378f8 req-29142090-51e1-4984-8ff5-95597d1601b4 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Received event network-changed-7fb046f4-e470-4f6d-8fa3-73f3b0157020 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 933.112823] env[62753]: DEBUG nova.compute.manager [req-770f2557-6f96-4d48-8621-64bb597378f8 req-29142090-51e1-4984-8ff5-95597d1601b4 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Refreshing instance network info cache due to event network-changed-7fb046f4-e470-4f6d-8fa3-73f3b0157020. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 933.113387] env[62753]: DEBUG oslo_concurrency.lockutils [req-770f2557-6f96-4d48-8621-64bb597378f8 req-29142090-51e1-4984-8ff5-95597d1601b4 service nova] Acquiring lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.113825] env[62753]: DEBUG oslo_concurrency.lockutils [req-770f2557-6f96-4d48-8621-64bb597378f8 req-29142090-51e1-4984-8ff5-95597d1601b4 service nova] Acquired lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.114272] env[62753]: DEBUG nova.network.neutron [req-770f2557-6f96-4d48-8621-64bb597378f8 req-29142090-51e1-4984-8ff5-95597d1601b4 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Refreshing network info cache for port 7fb046f4-e470-4f6d-8fa3-73f3b0157020 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 933.299014] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332421, 'name': CreateVM_Task, 'duration_secs': 0.508708} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.299481] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 933.300490] env[62753]: DEBUG oslo_concurrency.lockutils [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 933.300852] env[62753]: DEBUG oslo_concurrency.lockutils [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 933.301361] env[62753]: DEBUG oslo_concurrency.lockutils [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 933.301755] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fcd4bbbd-032a-444f-82ef-cc8d404466d5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.311521] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 933.311521] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529fda4e-0362-5cf3-60ce-6a2d65418a6a" [ 933.311521] env[62753]: _type = "Task" [ 933.311521] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.326937] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529fda4e-0362-5cf3-60ce-6a2d65418a6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.337449] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dc994f-44b2-528c-25b0-db9c56975c73, 'name': SearchDatastore_Task, 'duration_secs': 0.014634} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.338335] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a15d306-3e59-432f-8a7b-6ca96ddd7324 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.347879] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 933.347879] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5249df44-76b4-0e5c-c2e0-bdabad339f60" [ 933.347879] env[62753]: _type = "Task" [ 933.347879] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.358558] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5249df44-76b4-0e5c-c2e0-bdabad339f60, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.423321] env[62753]: DEBUG nova.scheduler.client.report [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 933.536726] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332423, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.611915] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "c3679c66-2763-4a04-8d14-e103bd952798" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 933.611915] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "c3679c66-2763-4a04-8d14-e103bd952798" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.633984] env[62753]: DEBUG nova.network.neutron [req-89045458-3d50-40f7-b895-769e3c5f671d req-0fda8cb1-0427-4923-8194-a14441e151d3 service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Updated VIF entry in instance network info cache for port 6ea2368a-1e67-4403-9a24-10736cc87080. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 933.634453] env[62753]: DEBUG nova.network.neutron [req-89045458-3d50-40f7-b895-769e3c5f671d req-0fda8cb1-0427-4923-8194-a14441e151d3 service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Updating instance_info_cache with network_info: [{"id": "167ccdaa-b799-422b-b031-ea36de9e571a", "address": "fa:16:3e:7f:2d:56", "network": {"id": "61486b9e-1554-4e34-990f-38a03833f61f", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-855522933", "subnets": [{"cidr": "192.168.128.0/24", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.39", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4b01e80a17e843a98ee662157230ae17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "6b8137fc-f23d-49b1-b19c-3123a5588f34", "external-id": "nsx-vlan-transportzone-709", "segmentation_id": 709, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap167ccdaa-b7", "ovs_interfaceid": "167ccdaa-b799-422b-b031-ea36de9e571a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "6ea2368a-1e67-4403-9a24-10736cc87080", "address": "fa:16:3e:24:f0:ea", "network": {"id": "c2c606e7-dc7b-4aa4-9abc-b5299f6b4660", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-455294558", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.127", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4b01e80a17e843a98ee662157230ae17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e839c46-1ae9-43b7-9518-8f18f48100dd", "external-id": "nsx-vlan-transportzone-666", "segmentation_id": 666, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ea2368a-1e", "ovs_interfaceid": "6ea2368a-1e67-4403-9a24-10736cc87080", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.799584] env[62753]: DEBUG nova.network.neutron [req-9a784af9-536f-4f8c-b17a-691dc33880e4 req-998f8703-10c2-4bf4-ad6f-27670c7d7fb1 service nova] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Updated VIF entry in instance network info cache for port f433dba2-2032-4957-8ba6-06568d931ccc. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 933.799584] env[62753]: DEBUG nova.network.neutron [req-9a784af9-536f-4f8c-b17a-691dc33880e4 req-998f8703-10c2-4bf4-ad6f-27670c7d7fb1 service nova] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Updating instance_info_cache with network_info: [{"id": "f433dba2-2032-4957-8ba6-06568d931ccc", "address": "fa:16:3e:97:cc:94", "network": {"id": "35ed0ec0-821f-45b6-bc2d-015626bfc0a2", "bridge": "br-int", "label": "tempest-ServersTestJSON-444726172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5636da907ab343f9b42dbbd903d32283", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf433dba2-20", "ovs_interfaceid": "f433dba2-2032-4957-8ba6-06568d931ccc", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.822889] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529fda4e-0362-5cf3-60ce-6a2d65418a6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.860297] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5249df44-76b4-0e5c-c2e0-bdabad339f60, 'name': SearchDatastore_Task, 'duration_secs': 0.020526} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 933.860570] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 933.860838] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 8066aeb2-7be4-46e3-aba6-124fef342a49/8066aeb2-7be4-46e3-aba6-124fef342a49.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 933.861128] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-b57e4418-1d81-45ea-bd1c-9288ae89d355 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 933.866782] env[62753]: DEBUG nova.network.neutron [req-770f2557-6f96-4d48-8621-64bb597378f8 req-29142090-51e1-4984-8ff5-95597d1601b4 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Updated VIF entry in instance network info cache for port 7fb046f4-e470-4f6d-8fa3-73f3b0157020. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 933.867164] env[62753]: DEBUG nova.network.neutron [req-770f2557-6f96-4d48-8621-64bb597378f8 req-29142090-51e1-4984-8ff5-95597d1601b4 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Updating instance_info_cache with network_info: [{"id": "7fb046f4-e470-4f6d-8fa3-73f3b0157020", "address": "fa:16:3e:6f:26:2c", "network": {"id": "fc646f7f-70ba-4671-8476-3b0ece1e9041", "bridge": null, "label": "tempest-AttachVolumeShelveTestJSON-853174775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d0eb0b3ba504f59a85cb021da2b47b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap7fb046f4-e4", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 933.870009] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 933.870009] env[62753]: value = "task-1332424" [ 933.870009] env[62753]: _type = "Task" [ 933.870009] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 933.881320] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332424, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 933.929101] env[62753]: DEBUG oslo_concurrency.lockutils [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.869s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 933.932576] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.610s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 933.953743] env[62753]: INFO nova.scheduler.client.report [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Deleted allocations for instance 50e1c1a1-7b8f-49cd-932d-03c920209634 [ 934.035530] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332423, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.085354] env[62753]: DEBUG nova.compute.manager [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 934.115036] env[62753]: DEBUG nova.compute.manager [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 934.120048] env[62753]: DEBUG nova.virt.hardware [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 934.120338] env[62753]: DEBUG nova.virt.hardware [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 934.120503] env[62753]: DEBUG nova.virt.hardware [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 934.120690] env[62753]: DEBUG nova.virt.hardware [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 934.120837] env[62753]: DEBUG nova.virt.hardware [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 934.120997] env[62753]: DEBUG nova.virt.hardware [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 934.121248] env[62753]: DEBUG nova.virt.hardware [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 934.121417] env[62753]: DEBUG nova.virt.hardware [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 934.121594] env[62753]: DEBUG nova.virt.hardware [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 934.121763] env[62753]: DEBUG nova.virt.hardware [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 934.121943] env[62753]: DEBUG nova.virt.hardware [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 934.122916] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e37ede38-f385-4112-8e87-8d44cd64aff7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.134129] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f72f4980-d1e3-4f08-903b-535d669f1a2a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.139910] env[62753]: DEBUG oslo_concurrency.lockutils [req-89045458-3d50-40f7-b895-769e3c5f671d req-0fda8cb1-0427-4923-8194-a14441e151d3 service nova] Releasing lock "refresh_cache-f8470f38-0150-41c6-a3ee-aabb5e68d98b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.302506] env[62753]: DEBUG oslo_concurrency.lockutils [req-9a784af9-536f-4f8c-b17a-691dc33880e4 req-998f8703-10c2-4bf4-ad6f-27670c7d7fb1 service nova] Releasing lock "refresh_cache-8066aeb2-7be4-46e3-aba6-124fef342a49" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.325914] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529fda4e-0362-5cf3-60ce-6a2d65418a6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.373160] env[62753]: DEBUG oslo_concurrency.lockutils [req-770f2557-6f96-4d48-8621-64bb597378f8 req-29142090-51e1-4984-8ff5-95597d1601b4 service nova] Releasing lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 934.383334] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332424, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.49074} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.383334] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 8066aeb2-7be4-46e3-aba6-124fef342a49/8066aeb2-7be4-46e3-aba6-124fef342a49.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 934.383334] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 934.383334] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1b3cd953-545f-4d3d-a320-458c3a0311fa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.393644] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 934.393644] env[62753]: value = "task-1332425" [ 934.393644] env[62753]: _type = "Task" [ 934.393644] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.404030] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332425, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.439197] env[62753]: INFO nova.compute.claims [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 934.463118] env[62753]: DEBUG oslo_concurrency.lockutils [None req-71038ac2-23f4-4235-bc77-6ba035baa57d tempest-ServerShowV247Test-484510260 tempest-ServerShowV247Test-484510260-project-member] Lock "50e1c1a1-7b8f-49cd-932d-03c920209634" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.257s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.534040] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332423, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.556389] env[62753]: DEBUG nova.network.neutron [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Successfully updated port: 800db109-09e3-45a4-8753-20d9911c0677 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 934.642238] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.823390] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529fda4e-0362-5cf3-60ce-6a2d65418a6a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.904372] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332425, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.065548} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 934.904780] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 934.905638] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee276bac-acb6-4daf-9a03-5031c2d4b610 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.931472] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Reconfiguring VM instance instance-0000004f to attach disk [datastore2] 8066aeb2-7be4-46e3-aba6-124fef342a49/8066aeb2-7be4-46e3-aba6-124fef342a49.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 934.931914] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f20fe56a-9085-43c6-a478-75e16751901f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 934.949669] env[62753]: INFO nova.compute.resource_tracker [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Updating resource usage from migration 3c736453-0224-4471-9d77-b23fddf9618f [ 934.960806] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 934.960806] env[62753]: value = "task-1332426" [ 934.960806] env[62753]: _type = "Task" [ 934.960806] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 934.971546] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332426, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 934.976733] env[62753]: DEBUG nova.compute.manager [req-0f4e203c-73ff-4440-aa14-a43452cd593f req-38869f34-3856-4add-8da4-8daf85cdb942 service nova] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Received event network-vif-plugged-800db109-09e3-45a4-8753-20d9911c0677 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.977117] env[62753]: DEBUG oslo_concurrency.lockutils [req-0f4e203c-73ff-4440-aa14-a43452cd593f req-38869f34-3856-4add-8da4-8daf85cdb942 service nova] Acquiring lock "628239e8-a447-4183-bed9-5550e53fda2b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 934.977369] env[62753]: DEBUG oslo_concurrency.lockutils [req-0f4e203c-73ff-4440-aa14-a43452cd593f req-38869f34-3856-4add-8da4-8daf85cdb942 service nova] Lock "628239e8-a447-4183-bed9-5550e53fda2b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 934.977490] env[62753]: DEBUG oslo_concurrency.lockutils [req-0f4e203c-73ff-4440-aa14-a43452cd593f req-38869f34-3856-4add-8da4-8daf85cdb942 service nova] Lock "628239e8-a447-4183-bed9-5550e53fda2b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 934.977748] env[62753]: DEBUG nova.compute.manager [req-0f4e203c-73ff-4440-aa14-a43452cd593f req-38869f34-3856-4add-8da4-8daf85cdb942 service nova] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] No waiting events found dispatching network-vif-plugged-800db109-09e3-45a4-8753-20d9911c0677 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 934.977879] env[62753]: WARNING nova.compute.manager [req-0f4e203c-73ff-4440-aa14-a43452cd593f req-38869f34-3856-4add-8da4-8daf85cdb942 service nova] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Received unexpected event network-vif-plugged-800db109-09e3-45a4-8753-20d9911c0677 for instance with vm_state building and task_state spawning. [ 934.978109] env[62753]: DEBUG nova.compute.manager [req-0f4e203c-73ff-4440-aa14-a43452cd593f req-38869f34-3856-4add-8da4-8daf85cdb942 service nova] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Received event network-changed-800db109-09e3-45a4-8753-20d9911c0677 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 934.978311] env[62753]: DEBUG nova.compute.manager [req-0f4e203c-73ff-4440-aa14-a43452cd593f req-38869f34-3856-4add-8da4-8daf85cdb942 service nova] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Refreshing instance network info cache due to event network-changed-800db109-09e3-45a4-8753-20d9911c0677. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 934.978533] env[62753]: DEBUG oslo_concurrency.lockutils [req-0f4e203c-73ff-4440-aa14-a43452cd593f req-38869f34-3856-4add-8da4-8daf85cdb942 service nova] Acquiring lock "refresh_cache-628239e8-a447-4183-bed9-5550e53fda2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 934.978773] env[62753]: DEBUG oslo_concurrency.lockutils [req-0f4e203c-73ff-4440-aa14-a43452cd593f req-38869f34-3856-4add-8da4-8daf85cdb942 service nova] Acquired lock "refresh_cache-628239e8-a447-4183-bed9-5550e53fda2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 934.979053] env[62753]: DEBUG nova.network.neutron [req-0f4e203c-73ff-4440-aa14-a43452cd593f req-38869f34-3856-4add-8da4-8daf85cdb942 service nova] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Refreshing network info cache for port 800db109-09e3-45a4-8753-20d9911c0677 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 935.036618] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332423, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.058967] env[62753]: DEBUG oslo_concurrency.lockutils [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "refresh_cache-628239e8-a447-4183-bed9-5550e53fda2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.183280] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b0721430-fd55-499b-bcd2-8e7465ea9038 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.205655] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d73e2d99-b423-4acf-91c6-80f035868f41 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.240142] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c1be9fd-ad02-47ea-a068-8fd489e8cc42 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.249613] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac3b2c5e-eff2-452f-bcac-30ae37e6bc9f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.266876] env[62753]: DEBUG nova.compute.provider_tree [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 935.324184] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529fda4e-0362-5cf3-60ce-6a2d65418a6a, 'name': SearchDatastore_Task, 'duration_secs': 1.563949} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.324494] env[62753]: DEBUG oslo_concurrency.lockutils [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.324747] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 935.325010] env[62753]: DEBUG oslo_concurrency.lockutils [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 935.325163] env[62753]: DEBUG oslo_concurrency.lockutils [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 935.325361] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 935.325663] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-175a6363-8d05-463f-b27c-1c448ca98bbb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.335397] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 935.335397] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 935.336165] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-57221558-4374-4ed0-b6d7-2e23b040616e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.342734] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 935.342734] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]524a0b18-6f84-d92e-9619-a87651b1c5bc" [ 935.342734] env[62753]: _type = "Task" [ 935.342734] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.355106] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]524a0b18-6f84-d92e-9619-a87651b1c5bc, 'name': SearchDatastore_Task, 'duration_secs': 0.010123} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.355982] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8d44efbe-fcf2-4070-a165-3863d05739ec {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.361865] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 935.361865] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dffdae-97ba-5d9a-7ffe-78dc885e4837" [ 935.361865] env[62753]: _type = "Task" [ 935.361865] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.370459] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dffdae-97ba-5d9a-7ffe-78dc885e4837, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.471601] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332426, 'name': ReconfigVM_Task, 'duration_secs': 0.337348} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.471996] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Reconfigured VM instance instance-0000004f to attach disk [datastore2] 8066aeb2-7be4-46e3-aba6-124fef342a49/8066aeb2-7be4-46e3-aba6-124fef342a49.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 935.472765] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e91a9d27-c6e3-4897-8203-81500a1ef823 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.481146] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 935.481146] env[62753]: value = "task-1332427" [ 935.481146] env[62753]: _type = "Task" [ 935.481146] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.493134] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332427, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.517345] env[62753]: DEBUG nova.network.neutron [req-0f4e203c-73ff-4440-aa14-a43452cd593f req-38869f34-3856-4add-8da4-8daf85cdb942 service nova] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 935.536720] env[62753]: DEBUG oslo_vmware.api [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332423, 'name': DeleteDatastoreFile_Task, 'duration_secs': 2.046029} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.537023] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 935.537417] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 935.537417] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 935.562568] env[62753]: INFO nova.scheduler.client.report [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Deleted allocations for instance 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5 [ 935.603121] env[62753]: DEBUG nova.network.neutron [req-0f4e203c-73ff-4440-aa14-a43452cd593f req-38869f34-3856-4add-8da4-8daf85cdb942 service nova] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 935.770551] env[62753]: DEBUG nova.scheduler.client.report [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 935.874631] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dffdae-97ba-5d9a-7ffe-78dc885e4837, 'name': SearchDatastore_Task, 'duration_secs': 0.009492} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.874923] env[62753]: DEBUG oslo_concurrency.lockutils [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 935.875443] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] f8470f38-0150-41c6-a3ee-aabb5e68d98b/f8470f38-0150-41c6-a3ee-aabb5e68d98b.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 935.875564] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8debfe8f-6c33-411a-ab0d-ef8c7ac0f7ca {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 935.882503] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 935.882503] env[62753]: value = "task-1332428" [ 935.882503] env[62753]: _type = "Task" [ 935.882503] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 935.891776] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332428, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 935.991948] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332427, 'name': Rename_Task, 'duration_secs': 0.151162} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 935.992285] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 935.992608] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-51743c46-92b3-4aa0-9432-ce59783dd2b7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.001454] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 936.001454] env[62753]: value = "task-1332429" [ 936.001454] env[62753]: _type = "Task" [ 936.001454] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.010717] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332429, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.067282] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 936.106487] env[62753]: DEBUG oslo_concurrency.lockutils [req-0f4e203c-73ff-4440-aa14-a43452cd593f req-38869f34-3856-4add-8da4-8daf85cdb942 service nova] Releasing lock "refresh_cache-628239e8-a447-4183-bed9-5550e53fda2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.106937] env[62753]: DEBUG oslo_concurrency.lockutils [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "refresh_cache-628239e8-a447-4183-bed9-5550e53fda2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.107151] env[62753]: DEBUG nova.network.neutron [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 936.276818] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.344s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 936.277129] env[62753]: INFO nova.compute.manager [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Migrating [ 936.277171] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "compute-rpcapi-router" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 936.277336] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "compute-rpcapi-router" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 936.278794] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.232s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 936.279152] env[62753]: DEBUG nova.objects.instance [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Lazy-loading 'resources' on Instance uuid c5dae9a3-052b-4f4c-ac79-84aeaded457a {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 936.400991] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332428, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.515639] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332429, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 936.650180] env[62753]: DEBUG nova.network.neutron [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 936.782481] env[62753]: INFO nova.compute.rpcapi [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Automatically selected compute RPC version 6.3 from minimum service version 67 [ 936.783358] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "compute-rpcapi-router" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 936.850028] env[62753]: DEBUG nova.network.neutron [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Updating instance_info_cache with network_info: [{"id": "800db109-09e3-45a4-8753-20d9911c0677", "address": "fa:16:3e:0e:c9:9f", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap800db109-09", "ovs_interfaceid": "800db109-09e3-45a4-8753-20d9911c0677", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 936.897377] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332428, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.687543} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 936.897664] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] f8470f38-0150-41c6-a3ee-aabb5e68d98b/f8470f38-0150-41c6-a3ee-aabb5e68d98b.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 936.897888] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 936.898271] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-5cd76dd8-ade2-4154-b206-0bd462deba19 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 936.909887] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 936.909887] env[62753]: value = "task-1332430" [ 936.909887] env[62753]: _type = "Task" [ 936.909887] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 936.924149] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332430, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.014254] env[62753]: DEBUG oslo_vmware.api [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332429, 'name': PowerOnVM_Task, 'duration_secs': 0.83319} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.015652] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 937.015875] env[62753]: INFO nova.compute.manager [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Took 7.15 seconds to spawn the instance on the hypervisor. [ 937.016119] env[62753]: DEBUG nova.compute.manager [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 937.016916] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7f42528-3a7b-49db-94a4-83d0c18f02e0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.020060] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c54c4973-a36d-443f-beff-b58f048febb4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.030049] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f98afb3-840a-4b51-bddb-9d504a92bf47 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.069724] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40bf8c94-32c3-4e3d-929a-c5a04c64e6bd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.079476] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc34a5c5-2670-4b1f-980a-c7a8ba65bef7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.095949] env[62753]: DEBUG nova.compute.provider_tree [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 937.307564] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "refresh_cache-a401909c-5978-492b-a4fa-b17d38a5af44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 937.307564] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "refresh_cache-a401909c-5978-492b-a4fa-b17d38a5af44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 937.307564] env[62753]: DEBUG nova.network.neutron [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 937.358516] env[62753]: DEBUG oslo_concurrency.lockutils [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "refresh_cache-628239e8-a447-4183-bed9-5550e53fda2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 937.359268] env[62753]: DEBUG nova.compute.manager [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Instance network_info: |[{"id": "800db109-09e3-45a4-8753-20d9911c0677", "address": "fa:16:3e:0e:c9:9f", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap800db109-09", "ovs_interfaceid": "800db109-09e3-45a4-8753-20d9911c0677", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 937.359621] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:c9:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a10c88d7-d13f-44fd-acee-7a734eb5f56a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '800db109-09e3-45a4-8753-20d9911c0677', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 937.368909] env[62753]: DEBUG oslo.service.loopingcall [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 937.369162] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 937.369408] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3e806444-0995-4e39-b972-30f8534807b8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.392603] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 937.392603] env[62753]: value = "task-1332431" [ 937.392603] env[62753]: _type = "Task" [ 937.392603] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.402116] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332431, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.419767] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332430, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.159062} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 937.420068] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 937.420969] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e57890a-3d50-4304-a84a-8637649c7c99 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.448121] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Reconfiguring VM instance instance-0000004e to attach disk [datastore1] f8470f38-0150-41c6-a3ee-aabb5e68d98b/f8470f38-0150-41c6-a3ee-aabb5e68d98b.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 937.448476] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-af20ecd4-25a4-4ffe-8873-7e1ee7a9af38 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 937.472553] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 937.472553] env[62753]: value = "task-1332432" [ 937.472553] env[62753]: _type = "Task" [ 937.472553] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 937.484280] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332432, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.545912] env[62753]: INFO nova.compute.manager [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Took 14.28 seconds to build instance. [ 937.602317] env[62753]: DEBUG nova.scheduler.client.report [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 937.909292] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332431, 'name': CreateVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 937.988700] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332432, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.047745] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e2199b9-9e7d-478b-a64a-502b9c09b0a3 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "8066aeb2-7be4-46e3-aba6-124fef342a49" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.795s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.071530] env[62753]: DEBUG nova.network.neutron [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Updating instance_info_cache with network_info: [{"id": "7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36", "address": "fa:16:3e:52:63:e3", "network": {"id": "2823647f-d32d-4a1c-9e02-7577016a260e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1594498922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6c8a6bc946a469fa85cc3ab80d7333a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7df5849e-e6", "ovs_interfaceid": "7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 938.105825] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.827s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.108382] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.757s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.108513] env[62753]: DEBUG nova.objects.instance [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lazy-loading 'resources' on Instance uuid b45bfcb4-b2cd-468c-9595-62c155d520fb {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 938.131784] env[62753]: INFO nova.scheduler.client.report [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Deleted allocations for instance c5dae9a3-052b-4f4c-ac79-84aeaded457a [ 938.407065] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332431, 'name': CreateVM_Task, 'duration_secs': 0.626806} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.407499] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 938.407958] env[62753]: DEBUG oslo_concurrency.lockutils [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.408144] env[62753]: DEBUG oslo_concurrency.lockutils [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.408838] env[62753]: DEBUG oslo_concurrency.lockutils [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 938.409129] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4b163671-3a24-4f22-8102-0a60c414a416 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.414615] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 938.414615] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52445bd6-5c32-a040-effb-5ee5a81e2830" [ 938.414615] env[62753]: _type = "Task" [ 938.414615] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.423223] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52445bd6-5c32-a040-effb-5ee5a81e2830, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.484272] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332432, 'name': ReconfigVM_Task, 'duration_secs': 0.727632} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.484630] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Reconfigured VM instance instance-0000004e to attach disk [datastore1] f8470f38-0150-41c6-a3ee-aabb5e68d98b/f8470f38-0150-41c6-a3ee-aabb5e68d98b.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 938.485332] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7ed0804b-ce5f-4685-b58e-df7f21106b28 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.492553] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 938.492553] env[62753]: value = "task-1332433" [ 938.492553] env[62753]: _type = "Task" [ 938.492553] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.502901] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332433, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.566717] env[62753]: DEBUG oslo_vmware.rw_handles [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52431f05-8fa3-faad-6709-e9d66ebe5e64/disk-0.vmdk. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 938.567632] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd94fffd-bee5-457f-b0fe-26e676e2d129 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.574831] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "refresh_cache-a401909c-5978-492b-a4fa-b17d38a5af44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.576284] env[62753]: DEBUG oslo_vmware.rw_handles [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52431f05-8fa3-faad-6709-e9d66ebe5e64/disk-0.vmdk is in state: ready. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 938.576449] env[62753]: ERROR oslo_vmware.rw_handles [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52431f05-8fa3-faad-6709-e9d66ebe5e64/disk-0.vmdk due to incomplete transfer. [ 938.576683] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-9836a71a-ad1e-4d86-9ae6-5b7985d348c3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.585301] env[62753]: DEBUG oslo_vmware.rw_handles [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52431f05-8fa3-faad-6709-e9d66ebe5e64/disk-0.vmdk. {{(pid=62753) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 938.585372] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Uploaded image 673e56af-bdea-458e-b52c-3abfd1149b15 to the Glance image server {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 938.588008] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Destroying the VM {{(pid=62753) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 938.588141] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-fa6d572f-5347-4f13-93d0-f36d9bb8ed8f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.595484] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 938.595484] env[62753]: value = "task-1332434" [ 938.595484] env[62753]: _type = "Task" [ 938.595484] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.605211] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332434, 'name': Destroy_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 938.642965] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b9020a75-ac7a-4f01-976f-6ae24bd723c7 tempest-ServersTestBootFromVolume-1780867481 tempest-ServersTestBootFromVolume-1780867481-project-member] Lock "c5dae9a3-052b-4f4c-ac79-84aeaded457a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.865s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 938.784732] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e844d03-c611-4969-9080-d63eea4777b9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.793876] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c87cf27b-b1d0-4aeb-9eea-ee432b9f07b9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.829446] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bf65114d-4de4-4b9a-af55-153811a6b9ec {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.835870] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "abd5ec6e-7b12-433b-9009-f9108ab4e318" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 938.836106] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "abd5ec6e-7b12-433b-9009-f9108ab4e318" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 938.843838] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1d2acc3-53a1-4683-95da-274085911309 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.860576] env[62753]: DEBUG nova.compute.provider_tree [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 938.927011] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52445bd6-5c32-a040-effb-5ee5a81e2830, 'name': SearchDatastore_Task, 'duration_secs': 0.028412} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 938.927384] env[62753]: DEBUG oslo_concurrency.lockutils [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 938.927691] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 938.927994] env[62753]: DEBUG oslo_concurrency.lockutils [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 938.928191] env[62753]: DEBUG oslo_concurrency.lockutils [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 938.928416] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 938.928745] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-82832528-a970-4c41-a3c3-7b0bbe542f56 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.939185] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 938.939444] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 938.940296] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0dcd06dd-ff55-4835-ac81-cfa67f580600 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 938.946756] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 938.946756] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52410612-0294-06a6-7be4-8c4a3bebab7d" [ 938.946756] env[62753]: _type = "Task" [ 938.946756] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 938.956626] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52410612-0294-06a6-7be4-8c4a3bebab7d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.005885] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332433, 'name': Rename_Task, 'duration_secs': 0.380946} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.006174] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 939.006347] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-287b1d47-a760-4cc0-9e6f-3f8c3434d013 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.014013] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 939.014013] env[62753]: value = "task-1332435" [ 939.014013] env[62753]: _type = "Task" [ 939.014013] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.023803] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332435, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.106871] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332434, 'name': Destroy_Task} progress is 33%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.149211] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.339608] env[62753]: DEBUG nova.compute.manager [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 939.366019] env[62753]: DEBUG nova.scheduler.client.report [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 939.465382] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52410612-0294-06a6-7be4-8c4a3bebab7d, 'name': SearchDatastore_Task, 'duration_secs': 0.019826} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.466655] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfe4a042-d41a-457f-86fc-ec29ad127bdb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.473829] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 939.473829] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f00e17-f1c0-42ef-650b-5ef85da12eb7" [ 939.473829] env[62753]: _type = "Task" [ 939.473829] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.485974] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f00e17-f1c0-42ef-650b-5ef85da12eb7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.527231] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332435, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.610107] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332434, 'name': Destroy_Task, 'duration_secs': 0.625832} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.611892] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Destroyed the VM [ 939.611892] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Deleting Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 939.611892] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-4616321e-525b-47c0-a1df-feac815d6474 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.621888] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 939.621888] env[62753]: value = "task-1332436" [ 939.621888] env[62753]: _type = "Task" [ 939.621888] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 939.631892] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332436, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 939.870887] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.762s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 939.873597] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.232s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 939.875208] env[62753]: INFO nova.compute.claims [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 939.879537] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 939.900261] env[62753]: INFO nova.scheduler.client.report [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Deleted allocations for instance b45bfcb4-b2cd-468c-9595-62c155d520fb [ 939.986319] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f00e17-f1c0-42ef-650b-5ef85da12eb7, 'name': SearchDatastore_Task, 'duration_secs': 0.013719} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 939.986724] env[62753]: DEBUG oslo_concurrency.lockutils [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 939.987283] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 628239e8-a447-4183-bed9-5550e53fda2b/628239e8-a447-4183-bed9-5550e53fda2b.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 939.988374] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-389f0498-1f56-442f-9451-b40cde4b64dd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 939.999465] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 939.999465] env[62753]: value = "task-1332437" [ 939.999465] env[62753]: _type = "Task" [ 939.999465] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.012367] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332437, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.030437] env[62753]: DEBUG oslo_vmware.api [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332435, 'name': PowerOnVM_Task, 'duration_secs': 0.53687} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.031195] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 940.031549] env[62753]: INFO nova.compute.manager [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Took 12.46 seconds to spawn the instance on the hypervisor. [ 940.031939] env[62753]: DEBUG nova.compute.manager [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 940.033275] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7b96cb2-6d4c-427f-ab41-afd3a19cc22a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.092673] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-056d2fc3-50d2-4d19-ae67-cf8e46205a40 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.114835] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Updating instance 'a401909c-5978-492b-a4fa-b17d38a5af44' progress to 0 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 940.133760] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332436, 'name': RemoveSnapshot_Task} progress is 97%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.410305] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9e531093-de3a-4ef8-947b-386b0f7f959b tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "b45bfcb4-b2cd-468c-9595-62c155d520fb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.131s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 940.514809] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332437, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.559267] env[62753]: INFO nova.compute.manager [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Took 17.37 seconds to build instance. [ 940.622219] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 940.622683] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-df6dcd5b-8ef0-4dcb-9810-41cb464a1fdb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 940.635192] env[62753]: DEBUG oslo_vmware.api [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332436, 'name': RemoveSnapshot_Task, 'duration_secs': 0.900763} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 940.636713] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Deleted Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 940.636713] env[62753]: INFO nova.compute.manager [None req-a74abff9-7fa2-4503-8482-25f74dec72b4 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Took 15.23 seconds to snapshot the instance on the hypervisor. [ 940.638660] env[62753]: DEBUG oslo_vmware.api [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 940.638660] env[62753]: value = "task-1332438" [ 940.638660] env[62753]: _type = "Task" [ 940.638660] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 940.650065] env[62753]: DEBUG oslo_vmware.api [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332438, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 940.702555] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "ff4b295e-2eb2-42dd-b353-08df755f04ea" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 940.702929] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "ff4b295e-2eb2-42dd-b353-08df755f04ea" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 940.902453] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquiring lock "f8470f38-0150-41c6-a3ee-aabb5e68d98b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.013593] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332437, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.659668} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.013593] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 628239e8-a447-4183-bed9-5550e53fda2b/628239e8-a447-4183-bed9-5550e53fda2b.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 941.013593] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 941.013823] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a62d1244-d19c-4c12-a502-9e40198af7ac {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.021737] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 941.021737] env[62753]: value = "task-1332439" [ 941.021737] env[62753]: _type = "Task" [ 941.021737] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.033685] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332439, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.061464] env[62753]: DEBUG oslo_concurrency.lockutils [None req-38800af2-e5b3-4179-8cda-7a90e1b55269 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "f8470f38-0150-41c6-a3ee-aabb5e68d98b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 18.882s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.065184] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "f8470f38-0150-41c6-a3ee-aabb5e68d98b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.162s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.065184] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquiring lock "f8470f38-0150-41c6-a3ee-aabb5e68d98b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 941.065184] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "f8470f38-0150-41c6-a3ee-aabb5e68d98b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 941.065365] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "f8470f38-0150-41c6-a3ee-aabb5e68d98b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 941.070173] env[62753]: INFO nova.compute.manager [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Terminating instance [ 941.072072] env[62753]: DEBUG nova.compute.manager [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 941.072280] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 941.073164] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1546a2ed-f788-445d-a270-245851ce3c64 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.087563] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 941.087856] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-860fbd26-f037-498e-826b-f11e428befc0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.094501] env[62753]: DEBUG oslo_vmware.api [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 941.094501] env[62753]: value = "task-1332440" [ 941.094501] env[62753]: _type = "Task" [ 941.094501] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.105351] env[62753]: DEBUG oslo_vmware.api [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332440, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.125032] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a7c8ba6-f66b-42f5-9296-9257dc373980 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.816393] env[62753]: DEBUG nova.compute.manager [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 941.829103] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35b35018-456a-4700-8291-092547b686ce {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.839016] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332439, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068032} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.846407] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 941.846750] env[62753]: DEBUG oslo_vmware.api [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332440, 'name': PowerOffVM_Task, 'duration_secs': 0.224925} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.846969] env[62753]: DEBUG oslo_vmware.api [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332438, 'name': PowerOffVM_Task, 'duration_secs': 0.292807} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 941.848194] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0c1222-8e42-493e-af0e-1f344d91e3fa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.850772] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 941.850975] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 941.851260] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 941.851445] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Updating instance 'a401909c-5978-492b-a4fa-b17d38a5af44' progress to 17 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 941.879890] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ab54fc05-b663-4c6d-898e-6a380b23055c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.883827] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b050b62-d97e-4238-9f27-4f6d58458d13 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.904523] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Reconfiguring VM instance instance-00000050 to attach disk [datastore2] 628239e8-a447-4183-bed9-5550e53fda2b/628239e8-a447-4183-bed9-5550e53fda2b.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 941.905386] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d7939bb9-ba47-419d-8b37-5b2b96d257aa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.922993] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6d11f5d6-40cc-4b5f-bc96-e70e95b539a5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 941.928322] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 941.928322] env[62753]: value = "task-1332442" [ 941.928322] env[62753]: _type = "Task" [ 941.928322] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 941.941994] env[62753]: DEBUG nova.compute.provider_tree [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 941.949214] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332442, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 941.998314] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 941.998571] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 941.998753] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Deleting the datastore file [datastore1] f8470f38-0150-41c6-a3ee-aabb5e68d98b {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 941.999056] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-15d7d7cd-b312-4d56-806d-076a17077db2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.007200] env[62753]: DEBUG oslo_vmware.api [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for the task: (returnval){ [ 942.007200] env[62753]: value = "task-1332443" [ 942.007200] env[62753]: _type = "Task" [ 942.007200] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.015975] env[62753]: DEBUG oslo_vmware.api [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332443, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.353008] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 942.385712] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 942.386024] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 942.386232] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 942.386422] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 942.386574] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 942.386723] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 942.386932] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 942.387109] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 942.387284] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 942.387456] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 942.387701] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 942.392841] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-69142c55-90f6-434d-b344-52c71a21bfa4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.410051] env[62753]: DEBUG oslo_vmware.api [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 942.410051] env[62753]: value = "task-1332444" [ 942.410051] env[62753]: _type = "Task" [ 942.410051] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.419812] env[62753]: DEBUG oslo_vmware.api [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332444, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.438477] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332442, 'name': ReconfigVM_Task, 'duration_secs': 0.304293} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.438743] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Reconfigured VM instance instance-00000050 to attach disk [datastore2] 628239e8-a447-4183-bed9-5550e53fda2b/628239e8-a447-4183-bed9-5550e53fda2b.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 942.439628] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-21beb381-97ba-43b7-9699-8750e5445b20 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.445416] env[62753]: DEBUG nova.scheduler.client.report [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 942.450142] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 942.450142] env[62753]: value = "task-1332445" [ 942.450142] env[62753]: _type = "Task" [ 942.450142] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.459471] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332445, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 942.519454] env[62753]: DEBUG oslo_vmware.api [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Task: {'id': task-1332443, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.407212} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.519732] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 942.519913] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 942.520110] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 942.520292] env[62753]: INFO nova.compute.manager [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Took 1.45 seconds to destroy the instance on the hypervisor. [ 942.520546] env[62753]: DEBUG oslo.service.loopingcall [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 942.520750] env[62753]: DEBUG nova.compute.manager [-] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 942.521212] env[62753]: DEBUG nova.network.neutron [-] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 942.827036] env[62753]: DEBUG nova.compute.manager [req-35637516-3fdb-4cf7-8bed-2479d8f1fbbd req-0ddf33c3-537d-4f11-a85f-ab56c87b03cb service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Received event network-vif-deleted-167ccdaa-b799-422b-b031-ea36de9e571a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 942.827628] env[62753]: INFO nova.compute.manager [req-35637516-3fdb-4cf7-8bed-2479d8f1fbbd req-0ddf33c3-537d-4f11-a85f-ab56c87b03cb service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Neutron deleted interface 167ccdaa-b799-422b-b031-ea36de9e571a; detaching it from the instance and deleting it from the info cache [ 942.830995] env[62753]: DEBUG nova.network.neutron [req-35637516-3fdb-4cf7-8bed-2479d8f1fbbd req-0ddf33c3-537d-4f11-a85f-ab56c87b03cb service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Updating instance_info_cache with network_info: [{"id": "6ea2368a-1e67-4403-9a24-10736cc87080", "address": "fa:16:3e:24:f0:ea", "network": {"id": "c2c606e7-dc7b-4aa4-9abc-b5299f6b4660", "bridge": "br-int", "label": "tempest-ServersTestMultiNic-455294558", "subnets": [{"cidr": "192.168.129.0/24", "dns": [], "gateway": {"address": "192.168.129.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.129.127", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.129.2"}}], "meta": {"injected": false, "tenant_id": "4b01e80a17e843a98ee662157230ae17", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "5e839c46-1ae9-43b7-9518-8f18f48100dd", "external-id": "nsx-vlan-transportzone-666", "segmentation_id": 666, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6ea2368a-1e", "ovs_interfaceid": "6ea2368a-1e67-4403-9a24-10736cc87080", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 942.920839] env[62753]: DEBUG oslo_vmware.api [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332444, 'name': ReconfigVM_Task, 'duration_secs': 0.217391} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.921299] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Updating instance 'a401909c-5978-492b-a4fa-b17d38a5af44' progress to 33 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 942.952337] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.079s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 942.952956] env[62753]: DEBUG nova.compute.manager [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 942.955855] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.889s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 942.956106] env[62753]: DEBUG nova.objects.instance [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lazy-loading 'resources' on Instance uuid 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 942.966434] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332445, 'name': Rename_Task, 'duration_secs': 0.145628} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 942.966742] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 942.966995] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-10eb7db0-e2b6-4631-a16e-c4c705af25b2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 942.975391] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 942.975391] env[62753]: value = "task-1332446" [ 942.975391] env[62753]: _type = "Task" [ 942.975391] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 942.985688] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332446, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.336058] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-8db58691-410e-4dd8-a861-2dc81f82e827 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.346259] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e19ef4e-9ec9-49d7-9e8d-f3385f0021e6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.377087] env[62753]: DEBUG nova.compute.manager [req-35637516-3fdb-4cf7-8bed-2479d8f1fbbd req-0ddf33c3-537d-4f11-a85f-ab56c87b03cb service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Detach interface failed, port_id=167ccdaa-b799-422b-b031-ea36de9e571a, reason: Instance f8470f38-0150-41c6-a3ee-aabb5e68d98b could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 943.428773] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 943.429042] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 943.429191] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 943.429369] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 943.429525] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 943.429672] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 943.429887] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 943.430056] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 943.430233] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 943.430401] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 943.430578] env[62753]: DEBUG nova.virt.hardware [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 943.435886] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Reconfiguring VM instance instance-0000004c to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 943.436190] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-92f9593d-5d61-4692-be1c-57c7b50c684b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.455294] env[62753]: DEBUG oslo_vmware.api [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 943.455294] env[62753]: value = "task-1332447" [ 943.455294] env[62753]: _type = "Task" [ 943.455294] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 943.465218] env[62753]: DEBUG nova.compute.utils [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 943.466999] env[62753]: DEBUG nova.objects.instance [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lazy-loading 'numa_topology' on Instance uuid 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 943.467893] env[62753]: DEBUG oslo_vmware.api [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332447, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.468377] env[62753]: DEBUG nova.compute.manager [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 943.468597] env[62753]: DEBUG nova.network.neutron [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 943.485823] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332446, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 943.507633] env[62753]: DEBUG nova.policy [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9931b40cfd7846038805c6a4caedaac0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b8e81660b30408c998e412f5fa81469', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 943.517023] env[62753]: DEBUG nova.network.neutron [-] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 943.765689] env[62753]: DEBUG nova.network.neutron [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Successfully created port: 12ca481a-0cb3-465e-b04c-802ae3ce6438 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 943.970668] env[62753]: DEBUG oslo_vmware.api [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332447, 'name': ReconfigVM_Task, 'duration_secs': 0.159704} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 943.970668] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Reconfigured VM instance instance-0000004c to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 943.970668] env[62753]: DEBUG nova.compute.manager [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 943.973289] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deaebb21-2f71-4df5-87a5-1bff7a12b586 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 943.977168] env[62753]: DEBUG nova.objects.base [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Object Instance<4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5> lazy-loaded attributes: resources,numa_topology {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 943.989165] env[62753]: DEBUG oslo_vmware.api [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332446, 'name': PowerOnVM_Task, 'duration_secs': 1.004503} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.000899] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 944.002182] env[62753]: INFO nova.compute.manager [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Took 9.92 seconds to spawn the instance on the hypervisor. [ 944.002182] env[62753]: DEBUG nova.compute.manager [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 944.009368] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Reconfiguring VM instance instance-0000004c to attach disk [datastore1] a401909c-5978-492b-a4fa-b17d38a5af44/a401909c-5978-492b-a4fa-b17d38a5af44.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 944.012708] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab83ea74-9bfe-4e8c-89c0-b55afbcf5533 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.015697] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 944.015969] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77367aa0-25d6-487b-93de-fbd42f148c6a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.030351] env[62753]: INFO nova.compute.manager [-] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Took 1.51 seconds to deallocate network for instance. [ 944.042715] env[62753]: DEBUG oslo_vmware.api [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 944.042715] env[62753]: value = "task-1332448" [ 944.042715] env[62753]: _type = "Task" [ 944.042715] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 944.056765] env[62753]: DEBUG oslo_vmware.api [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332448, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 944.215564] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b560f492-2fa1-40a8-83f1-f3f9750dbdbf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.226071] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8bbeca8f-3ee5-422b-8320-f837fe3258d0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.257565] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff967655-6154-4817-a46c-1af4179351d7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.266497] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec53089d-e157-45a4-89cc-a37ea23c1e46 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 944.280610] env[62753]: DEBUG nova.compute.provider_tree [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 944.544450] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 944.556032] env[62753]: INFO nova.compute.manager [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Took 19.49 seconds to build instance. [ 944.561557] env[62753]: DEBUG oslo_vmware.api [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332448, 'name': ReconfigVM_Task, 'duration_secs': 0.29823} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 944.561901] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Reconfigured VM instance instance-0000004c to attach disk [datastore1] a401909c-5978-492b-a4fa-b17d38a5af44/a401909c-5978-492b-a4fa-b17d38a5af44.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 944.562190] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Updating instance 'a401909c-5978-492b-a4fa-b17d38a5af44' progress to 50 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 944.784066] env[62753]: DEBUG nova.scheduler.client.report [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 944.898753] env[62753]: DEBUG nova.compute.manager [req-c4fc8be4-30cf-47a1-a0f3-b126fb754e44 req-b5a13c6c-8115-41c0-999f-84f176362fda service nova] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Received event network-vif-deleted-6ea2368a-1e67-4403-9a24-10736cc87080 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 944.983399] env[62753]: DEBUG nova.compute.manager [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 945.007017] env[62753]: DEBUG nova.virt.hardware [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 945.007310] env[62753]: DEBUG nova.virt.hardware [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 945.007475] env[62753]: DEBUG nova.virt.hardware [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 945.007666] env[62753]: DEBUG nova.virt.hardware [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 945.007815] env[62753]: DEBUG nova.virt.hardware [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 945.007967] env[62753]: DEBUG nova.virt.hardware [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 945.008231] env[62753]: DEBUG nova.virt.hardware [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 945.008401] env[62753]: DEBUG nova.virt.hardware [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 945.008574] env[62753]: DEBUG nova.virt.hardware [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 945.008739] env[62753]: DEBUG nova.virt.hardware [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 945.008915] env[62753]: DEBUG nova.virt.hardware [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 945.009700] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 945.010488] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8753d14b-63ff-4415-a877-fe931bc90db6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.013538] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 945.013725] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62753) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 945.020366] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ac81fc4-6125-4b76-ae80-0ac26026350a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.058370] env[62753]: DEBUG oslo_concurrency.lockutils [None req-922b0e7b-2e92-4daf-9976-3e574fefa3f2 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "628239e8-a447-4183-bed9-5550e53fda2b" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.004s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.068406] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e7b48bd-b2b6-4cb2-baec-bbf1ad7d7a23 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.096072] env[62753]: DEBUG oslo_concurrency.lockutils [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "628239e8-a447-4183-bed9-5550e53fda2b" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 945.096519] env[62753]: DEBUG oslo_concurrency.lockutils [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "628239e8-a447-4183-bed9-5550e53fda2b" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.096861] env[62753]: INFO nova.compute.manager [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Shelving [ 945.100316] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b7f4544-d932-4147-82ad-4d714f03e8c9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.119869] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Updating instance 'a401909c-5978-492b-a4fa-b17d38a5af44' progress to 67 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 945.277039] env[62753]: DEBUG nova.network.neutron [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Successfully updated port: 12ca481a-0cb3-465e-b04c-802ae3ce6438 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 945.291685] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.336s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.294901] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.415s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.296492] env[62753]: INFO nova.compute.claims [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 945.608984] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 945.609283] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4cbcff6b-05f6-41b1-8b63-0284a87cff54 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 945.617908] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 945.617908] env[62753]: value = "task-1332449" [ 945.617908] env[62753]: _type = "Task" [ 945.617908] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 945.631297] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332449, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 945.666677] env[62753]: DEBUG nova.network.neutron [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Port 7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36 binding to destination host cpu-1 is already ACTIVE {{(pid=62753) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 945.779835] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 945.780009] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 945.780226] env[62753]: DEBUG nova.network.neutron [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 945.809363] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd9668b8-e694-4dd0-9d54-0c0caa2be2a4 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 33.395s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 945.810329] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 6.661s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 945.810548] env[62753]: INFO nova.compute.manager [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Unshelving [ 946.015309] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.015309] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.015309] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 946.128253] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332449, 'name': PowerOffVM_Task, 'duration_secs': 0.246642} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 946.128585] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 946.129611] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc5e084d-2e36-42b7-a578-b2e9f2cee68e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.150677] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be4b9398-4b12-4d1a-8101-014c54f82200 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.313197] env[62753]: DEBUG nova.network.neutron [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 946.455247] env[62753]: DEBUG nova.network.neutron [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Updating instance_info_cache with network_info: [{"id": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "address": "fa:16:3e:57:5d:75", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12ca481a-0c", "ovs_interfaceid": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 946.498734] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a5dbe25-f013-4674-98ce-b43f20549c57 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.506328] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1f5a0ecf-34b8-41ca-9fc7-2df2a4681068 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.536874] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b6d8951-3ea6-4bfc-ba2d-d6aa516615a2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.544793] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d44e389f-e319-40ec-b1f6-d6ae04fba019 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.558043] env[62753]: DEBUG nova.compute.provider_tree [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 946.664020] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Creating Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 946.664020] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-6e834b69-075a-4b1b-9ef9-0c9fb4c1fe2b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.671455] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 946.671455] env[62753]: value = "task-1332450" [ 946.671455] env[62753]: _type = "Task" [ 946.671455] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 946.694386] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "a401909c-5978-492b-a4fa-b17d38a5af44-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.694386] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "a401909c-5978-492b-a4fa-b17d38a5af44-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.694386] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "a401909c-5978-492b-a4fa-b17d38a5af44-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.695589] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332450, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 946.836804] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.926025] env[62753]: DEBUG nova.compute.manager [req-ab05eac3-f3cb-4c67-b37e-f5efe9b11cd1 req-75597932-abe5-4d99-9a46-8e0180676ec3 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Received event network-vif-plugged-12ca481a-0cb3-465e-b04c-802ae3ce6438 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.926255] env[62753]: DEBUG oslo_concurrency.lockutils [req-ab05eac3-f3cb-4c67-b37e-f5efe9b11cd1 req-75597932-abe5-4d99-9a46-8e0180676ec3 service nova] Acquiring lock "c3679c66-2763-4a04-8d14-e103bd952798-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 946.926847] env[62753]: DEBUG oslo_concurrency.lockutils [req-ab05eac3-f3cb-4c67-b37e-f5efe9b11cd1 req-75597932-abe5-4d99-9a46-8e0180676ec3 service nova] Lock "c3679c66-2763-4a04-8d14-e103bd952798-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 946.927206] env[62753]: DEBUG oslo_concurrency.lockutils [req-ab05eac3-f3cb-4c67-b37e-f5efe9b11cd1 req-75597932-abe5-4d99-9a46-8e0180676ec3 service nova] Lock "c3679c66-2763-4a04-8d14-e103bd952798-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 946.927525] env[62753]: DEBUG nova.compute.manager [req-ab05eac3-f3cb-4c67-b37e-f5efe9b11cd1 req-75597932-abe5-4d99-9a46-8e0180676ec3 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] No waiting events found dispatching network-vif-plugged-12ca481a-0cb3-465e-b04c-802ae3ce6438 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 946.927624] env[62753]: WARNING nova.compute.manager [req-ab05eac3-f3cb-4c67-b37e-f5efe9b11cd1 req-75597932-abe5-4d99-9a46-8e0180676ec3 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Received unexpected event network-vif-plugged-12ca481a-0cb3-465e-b04c-802ae3ce6438 for instance with vm_state building and task_state spawning. [ 946.927794] env[62753]: DEBUG nova.compute.manager [req-ab05eac3-f3cb-4c67-b37e-f5efe9b11cd1 req-75597932-abe5-4d99-9a46-8e0180676ec3 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Received event network-changed-12ca481a-0cb3-465e-b04c-802ae3ce6438 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 946.927956] env[62753]: DEBUG nova.compute.manager [req-ab05eac3-f3cb-4c67-b37e-f5efe9b11cd1 req-75597932-abe5-4d99-9a46-8e0180676ec3 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Refreshing instance network info cache due to event network-changed-12ca481a-0cb3-465e-b04c-802ae3ce6438. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 946.928151] env[62753]: DEBUG oslo_concurrency.lockutils [req-ab05eac3-f3cb-4c67-b37e-f5efe9b11cd1 req-75597932-abe5-4d99-9a46-8e0180676ec3 service nova] Acquiring lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 946.958288] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 946.958615] env[62753]: DEBUG nova.compute.manager [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Instance network_info: |[{"id": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "address": "fa:16:3e:57:5d:75", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12ca481a-0c", "ovs_interfaceid": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 946.958935] env[62753]: DEBUG oslo_concurrency.lockutils [req-ab05eac3-f3cb-4c67-b37e-f5efe9b11cd1 req-75597932-abe5-4d99-9a46-8e0180676ec3 service nova] Acquired lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 946.959131] env[62753]: DEBUG nova.network.neutron [req-ab05eac3-f3cb-4c67-b37e-f5efe9b11cd1 req-75597932-abe5-4d99-9a46-8e0180676ec3 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Refreshing network info cache for port 12ca481a-0cb3-465e-b04c-802ae3ce6438 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 946.960402] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:57:5d:75', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd48f0ef6-34e5-44d4-8baf-4470ed96ce73', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '12ca481a-0cb3-465e-b04c-802ae3ce6438', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 946.968790] env[62753]: DEBUG oslo.service.loopingcall [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 946.970067] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 946.970309] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b4cfe671-be3a-46b7-a340-8f5e63b6bbac {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 946.991936] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 946.991936] env[62753]: value = "task-1332451" [ 946.991936] env[62753]: _type = "Task" [ 946.991936] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.000593] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332451, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.006426] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 947.010071] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 947.061524] env[62753]: DEBUG nova.scheduler.client.report [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 947.182760] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332450, 'name': CreateSnapshot_Task, 'duration_secs': 0.39459} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 947.183084] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Created Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 947.183942] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7651965e-712c-4d34-9c44-349fe8a741ac {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.501405] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332451, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.512701] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 947.566651] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.271s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 947.569036] env[62753]: DEBUG nova.compute.manager [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 947.569779] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.217s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 947.571756] env[62753]: INFO nova.compute.claims [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 947.685486] env[62753]: DEBUG nova.network.neutron [req-ab05eac3-f3cb-4c67-b37e-f5efe9b11cd1 req-75597932-abe5-4d99-9a46-8e0180676ec3 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Updated VIF entry in instance network info cache for port 12ca481a-0cb3-465e-b04c-802ae3ce6438. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 947.685870] env[62753]: DEBUG nova.network.neutron [req-ab05eac3-f3cb-4c67-b37e-f5efe9b11cd1 req-75597932-abe5-4d99-9a46-8e0180676ec3 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Updating instance_info_cache with network_info: [{"id": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "address": "fa:16:3e:57:5d:75", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12ca481a-0c", "ovs_interfaceid": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 947.703227] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Creating linked-clone VM from snapshot {{(pid=62753) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 947.706556] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-f9b97a2f-fa10-4c94-ac60-e6a5a951419d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 947.719199] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 947.719199] env[62753]: value = "task-1332452" [ 947.719199] env[62753]: _type = "Task" [ 947.719199] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 947.728316] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332452, 'name': CloneVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 947.763589] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "refresh_cache-a401909c-5978-492b-a4fa-b17d38a5af44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 947.763822] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "refresh_cache-a401909c-5978-492b-a4fa-b17d38a5af44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 947.764012] env[62753]: DEBUG nova.network.neutron [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 948.006790] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332451, 'name': CreateVM_Task, 'duration_secs': 0.62874} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.007663] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 948.007981] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.008272] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.008795] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 948.009448] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-849dd0cd-1535-4bfa-b669-c245fb5180df {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.016046] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 948.016046] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bb3d96-68f0-ff0c-17dc-4b255afceaa3" [ 948.016046] env[62753]: _type = "Task" [ 948.016046] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.028161] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bb3d96-68f0-ff0c-17dc-4b255afceaa3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.078131] env[62753]: DEBUG nova.compute.utils [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 948.081237] env[62753]: DEBUG nova.compute.manager [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 948.081742] env[62753]: DEBUG nova.network.neutron [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 948.133186] env[62753]: DEBUG nova.policy [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4dbc63095b0464fa0d2de86e3cf170e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5636da907ab343f9b42dbbd903d32283', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 948.189310] env[62753]: DEBUG oslo_concurrency.lockutils [req-ab05eac3-f3cb-4c67-b37e-f5efe9b11cd1 req-75597932-abe5-4d99-9a46-8e0180676ec3 service nova] Releasing lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.230332] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332452, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.501254] env[62753]: DEBUG nova.network.neutron [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Successfully created port: a98dff4a-a9af-4961-8146-fbd28735e7af {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 948.528690] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bb3d96-68f0-ff0c-17dc-4b255afceaa3, 'name': SearchDatastore_Task, 'duration_secs': 0.012003} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 948.529046] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 948.529297] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 948.529559] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 948.529903] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 948.529903] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 948.530172] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ac9a9cf0-ef56-4676-a956-47173146716e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.540133] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 948.540334] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 948.543800] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-161cc0bc-2680-4fb5-ab75-63a625308bc8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.549892] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 948.549892] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c2a42a-c4b7-e2ed-9fec-ce6d4506855d" [ 948.549892] env[62753]: _type = "Task" [ 948.549892] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 948.560581] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c2a42a-c4b7-e2ed-9fec-ce6d4506855d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.581733] env[62753]: DEBUG nova.compute.manager [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 948.610586] env[62753]: DEBUG nova.network.neutron [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Updating instance_info_cache with network_info: [{"id": "7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36", "address": "fa:16:3e:52:63:e3", "network": {"id": "2823647f-d32d-4a1c-9e02-7577016a260e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1594498922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6c8a6bc946a469fa85cc3ab80d7333a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7df5849e-e6", "ovs_interfaceid": "7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 948.734407] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332452, 'name': CloneVM_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 948.793040] env[62753]: DEBUG nova.compute.manager [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 948.793533] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ff7837f-3c00-4003-96c8-0a623359688a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.802427] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b06724ef-d96a-4553-a4e0-f34f3a32abc7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.815278] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46e14b82-28c0-4896-8580-013bc4046210 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.847143] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-261a6b87-e054-47b0-8c1a-1e8689e5a6da {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.856145] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dd107c4c-cc95-4820-9f73-c808173e28af {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 948.870950] env[62753]: DEBUG nova.compute.provider_tree [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 949.062585] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c2a42a-c4b7-e2ed-9fec-ce6d4506855d, 'name': SearchDatastore_Task, 'duration_secs': 0.011117} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.063406] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-74872575-0d33-42ec-af54-68019fdace4c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.070048] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 949.070048] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52fc21ee-bfd8-9833-1d50-7b91566a72ab" [ 949.070048] env[62753]: _type = "Task" [ 949.070048] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.077624] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52fc21ee-bfd8-9833-1d50-7b91566a72ab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.116523] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "refresh_cache-a401909c-5978-492b-a4fa-b17d38a5af44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.232414] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332452, 'name': CloneVM_Task, 'duration_secs': 1.086772} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.232700] env[62753]: INFO nova.virt.vmwareapi.vmops [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Created linked-clone VM from snapshot [ 949.233513] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f32437ab-f38d-407e-a7c9-884b01adfa58 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.240819] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Uploading image 1bf6f512-055c-454e-b20a-aa68af11c3df {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 949.262861] env[62753]: DEBUG oslo_vmware.rw_handles [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 949.262861] env[62753]: value = "vm-284631" [ 949.262861] env[62753]: _type = "VirtualMachine" [ 949.262861] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 949.263140] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-49c76c96-bcc0-40b8-8354-bb6bb79c4ce8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.271153] env[62753]: DEBUG oslo_vmware.rw_handles [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lease: (returnval){ [ 949.271153] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c5092f-8bc3-cd1d-117b-51f4903d08e0" [ 949.271153] env[62753]: _type = "HttpNfcLease" [ 949.271153] env[62753]: } obtained for exporting VM: (result){ [ 949.271153] env[62753]: value = "vm-284631" [ 949.271153] env[62753]: _type = "VirtualMachine" [ 949.271153] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 949.271476] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the lease: (returnval){ [ 949.271476] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c5092f-8bc3-cd1d-117b-51f4903d08e0" [ 949.271476] env[62753]: _type = "HttpNfcLease" [ 949.271476] env[62753]: } to be ready. {{(pid=62753) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 949.278048] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 949.278048] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c5092f-8bc3-cd1d-117b-51f4903d08e0" [ 949.278048] env[62753]: _type = "HttpNfcLease" [ 949.278048] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 949.319712] env[62753]: INFO nova.compute.manager [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] instance snapshotting [ 949.321034] env[62753]: DEBUG nova.objects.instance [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lazy-loading 'flavor' on Instance uuid c73b1ae0-dc98-47f7-babf-e96169384785 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.374185] env[62753]: DEBUG nova.scheduler.client.report [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 949.580236] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52fc21ee-bfd8-9833-1d50-7b91566a72ab, 'name': SearchDatastore_Task, 'duration_secs': 0.038134} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 949.580502] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 949.580769] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] c3679c66-2763-4a04-8d14-e103bd952798/c3679c66-2763-4a04-8d14-e103bd952798.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 949.581044] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f0902669-62b5-4608-9c26-e81ab836d5e7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.588273] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 949.588273] env[62753]: value = "task-1332454" [ 949.588273] env[62753]: _type = "Task" [ 949.588273] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 949.593452] env[62753]: DEBUG nova.compute.manager [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 949.598465] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332454, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 949.616749] env[62753]: DEBUG nova.virt.hardware [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 949.617017] env[62753]: DEBUG nova.virt.hardware [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 949.617273] env[62753]: DEBUG nova.virt.hardware [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 949.617465] env[62753]: DEBUG nova.virt.hardware [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 949.617616] env[62753]: DEBUG nova.virt.hardware [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 949.617767] env[62753]: DEBUG nova.virt.hardware [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 949.617980] env[62753]: DEBUG nova.virt.hardware [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 949.618170] env[62753]: DEBUG nova.virt.hardware [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 949.618375] env[62753]: DEBUG nova.virt.hardware [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 949.618579] env[62753]: DEBUG nova.virt.hardware [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 949.618750] env[62753]: DEBUG nova.virt.hardware [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 949.619608] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd65729e-2f26-4be1-8987-e3a5e351f3fa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.632507] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c19725e0-91ba-4269-a530-51d3145317e1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.643880] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd45bf63-974f-4f3d-a26c-21ef05256edf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.672160] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d424421b-b174-405f-ab55-ab73ad08d33a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.680247] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Updating instance 'a401909c-5978-492b-a4fa-b17d38a5af44' progress to 83 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 949.722459] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 949.722459] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.781496] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 949.781496] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c5092f-8bc3-cd1d-117b-51f4903d08e0" [ 949.781496] env[62753]: _type = "HttpNfcLease" [ 949.781496] env[62753]: } is ready. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 949.781924] env[62753]: DEBUG oslo_vmware.rw_handles [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 949.781924] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c5092f-8bc3-cd1d-117b-51f4903d08e0" [ 949.781924] env[62753]: _type = "HttpNfcLease" [ 949.781924] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 949.782653] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e972f25e-62f1-4762-ba15-ba60eeb333ea {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.792366] env[62753]: DEBUG oslo_vmware.rw_handles [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc73a9-2e4b-5918-e94e-09342a64ac74/disk-0.vmdk from lease info. {{(pid=62753) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 949.792572] env[62753]: DEBUG oslo_vmware.rw_handles [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Opening URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc73a9-2e4b-5918-e94e-09342a64ac74/disk-0.vmdk for reading. {{(pid=62753) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 949.860866] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c884bd76-9745-4e4d-8e6e-75f23f9fab62 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.882323] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.312s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 949.882928] env[62753]: DEBUG nova.compute.manager [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 949.886048] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.342s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 949.886104] env[62753]: DEBUG nova.objects.instance [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lazy-loading 'resources' on Instance uuid f8470f38-0150-41c6-a3ee-aabb5e68d98b {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 949.887767] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c6e6f9b-5353-4e4c-91b4-a6aa6c71e695 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 949.923168] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-610a0fc9-3dfd-4d83-92e6-e237e5136a50 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.100426] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332454, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.161596] env[62753]: DEBUG nova.network.neutron [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Successfully updated port: a98dff4a-a9af-4961-8146-fbd28735e7af {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 950.187519] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 950.188899] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e5c27de8-069a-4ba8-a7aa-b8cdf7b111cc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.192030] env[62753]: DEBUG nova.compute.manager [req-6f2adc23-3dde-4e2b-861c-f871bada623c req-797816cd-326c-4835-8ef2-49221f35a6a8 service nova] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Received event network-vif-plugged-a98dff4a-a9af-4961-8146-fbd28735e7af {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 950.192254] env[62753]: DEBUG oslo_concurrency.lockutils [req-6f2adc23-3dde-4e2b-861c-f871bada623c req-797816cd-326c-4835-8ef2-49221f35a6a8 service nova] Acquiring lock "abd5ec6e-7b12-433b-9009-f9108ab4e318-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 950.192467] env[62753]: DEBUG oslo_concurrency.lockutils [req-6f2adc23-3dde-4e2b-861c-f871bada623c req-797816cd-326c-4835-8ef2-49221f35a6a8 service nova] Lock "abd5ec6e-7b12-433b-9009-f9108ab4e318-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 950.192638] env[62753]: DEBUG oslo_concurrency.lockutils [req-6f2adc23-3dde-4e2b-861c-f871bada623c req-797816cd-326c-4835-8ef2-49221f35a6a8 service nova] Lock "abd5ec6e-7b12-433b-9009-f9108ab4e318-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.192811] env[62753]: DEBUG nova.compute.manager [req-6f2adc23-3dde-4e2b-861c-f871bada623c req-797816cd-326c-4835-8ef2-49221f35a6a8 service nova] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] No waiting events found dispatching network-vif-plugged-a98dff4a-a9af-4961-8146-fbd28735e7af {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 950.192980] env[62753]: WARNING nova.compute.manager [req-6f2adc23-3dde-4e2b-861c-f871bada623c req-797816cd-326c-4835-8ef2-49221f35a6a8 service nova] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Received unexpected event network-vif-plugged-a98dff4a-a9af-4961-8146-fbd28735e7af for instance with vm_state building and task_state spawning. [ 950.202614] env[62753]: DEBUG oslo_vmware.api [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 950.202614] env[62753]: value = "task-1332455" [ 950.202614] env[62753]: _type = "Task" [ 950.202614] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.212718] env[62753]: DEBUG oslo_vmware.api [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332455, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.227076] env[62753]: DEBUG nova.compute.utils [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 950.396481] env[62753]: DEBUG nova.compute.utils [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 950.398405] env[62753]: DEBUG nova.compute.manager [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 950.399386] env[62753]: DEBUG nova.network.neutron [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 950.401823] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Creating Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 950.403015] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-bf366145-89f9-4328-b324-3057544430f2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.414520] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 950.414520] env[62753]: value = "task-1332456" [ 950.414520] env[62753]: _type = "Task" [ 950.414520] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.426016] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332456, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.449622] env[62753]: DEBUG nova.policy [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cbc9dc744e144ff28cf4c44fb043094f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e376836411ae4854965adf2923eaaffd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 950.602543] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332454, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.541608} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.605544] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] c3679c66-2763-4a04-8d14-e103bd952798/c3679c66-2763-4a04-8d14-e103bd952798.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 950.605805] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 950.606321] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1024c159-906d-44d1-9bc9-77ffc8d25d42 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.615345] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 950.615345] env[62753]: value = "task-1332457" [ 950.615345] env[62753]: _type = "Task" [ 950.615345] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 950.620715] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adb43ad1-963d-4045-a5d0-71a602bac346 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.634661] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7e5b2ff-8fd9-45f4-a614-0c142dc272de {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.637489] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332457, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 950.674057] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "refresh_cache-abd5ec6e-7b12-433b-9009-f9108ab4e318" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 950.674057] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "refresh_cache-abd5ec6e-7b12-433b-9009-f9108ab4e318" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 950.674204] env[62753]: DEBUG nova.network.neutron [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 950.677999] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-01fd050a-672a-445a-a0e3-adefb38c74c1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.690867] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a450d2df-c7fe-4f7e-81bc-6bad253d8a22 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 950.710952] env[62753]: DEBUG nova.compute.provider_tree [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 950.727079] env[62753]: DEBUG oslo_vmware.api [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332455, 'name': PowerOnVM_Task, 'duration_secs': 0.476528} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 950.728659] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 950.729240] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b77f9f03-4740-4d63-a79f-39a7a628d57d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Updating instance 'a401909c-5978-492b-a4fa-b17d38a5af44' progress to 100 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 950.734062] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.012s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 950.764810] env[62753]: DEBUG nova.network.neutron [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Successfully created port: 2003ddeb-12c2-4450-a45b-cc180b041d34 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 950.903307] env[62753]: DEBUG nova.compute.manager [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 950.932134] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332456, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.125367] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332457, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.127725} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.126040] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 951.126840] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab26ee32-14a9-4c0f-95e6-f17d9551bdd8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.150407] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Reconfiguring VM instance instance-00000051 to attach disk [datastore2] c3679c66-2763-4a04-8d14-e103bd952798/c3679c66-2763-4a04-8d14-e103bd952798.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 951.150767] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1cd2bbb1-f5da-4117-ab09-728845fbf5d3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.171870] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 951.171870] env[62753]: value = "task-1332458" [ 951.171870] env[62753]: _type = "Task" [ 951.171870] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.181818] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332458, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.210281] env[62753]: DEBUG nova.network.neutron [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 951.214350] env[62753]: DEBUG nova.scheduler.client.report [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 951.370180] env[62753]: DEBUG nova.network.neutron [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Updating instance_info_cache with network_info: [{"id": "a98dff4a-a9af-4961-8146-fbd28735e7af", "address": "fa:16:3e:e6:41:ee", "network": {"id": "35ed0ec0-821f-45b6-bc2d-015626bfc0a2", "bridge": "br-int", "label": "tempest-ServersTestJSON-444726172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5636da907ab343f9b42dbbd903d32283", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa98dff4a-a9", "ovs_interfaceid": "a98dff4a-a9af-4961-8146-fbd28735e7af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 951.430384] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332456, 'name': CreateSnapshot_Task, 'duration_secs': 0.72266} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 951.430829] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Created Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 951.431792] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62cbc962-2469-46ce-9cfa-9d3d3884c7fe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.682733] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332458, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.719860] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.834s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 951.722700] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 4.886s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.722700] env[62753]: DEBUG nova.objects.instance [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lazy-loading 'pci_requests' on Instance uuid 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 951.749076] env[62753]: INFO nova.scheduler.client.report [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Deleted allocations for instance f8470f38-0150-41c6-a3ee-aabb5e68d98b [ 951.797865] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 951.798178] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 951.798521] env[62753]: INFO nova.compute.manager [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Attaching volume 368def30-47d9-476e-8669-d2beafd3fba1 to /dev/sdb [ 951.834019] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-808b2a27-acd0-4814-8958-62e8cb112058 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.842934] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-feba8dc8-4ebd-4d28-9cee-0ff6555e6601 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.859684] env[62753]: DEBUG nova.virt.block_device [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Updating existing volume attachment record: 2796567e-b83b-4646-adad-7b0ec7615fc3 {{(pid=62753) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 951.873183] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "refresh_cache-abd5ec6e-7b12-433b-9009-f9108ab4e318" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 951.873701] env[62753]: DEBUG nova.compute.manager [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Instance network_info: |[{"id": "a98dff4a-a9af-4961-8146-fbd28735e7af", "address": "fa:16:3e:e6:41:ee", "network": {"id": "35ed0ec0-821f-45b6-bc2d-015626bfc0a2", "bridge": "br-int", "label": "tempest-ServersTestJSON-444726172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5636da907ab343f9b42dbbd903d32283", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa98dff4a-a9", "ovs_interfaceid": "a98dff4a-a9af-4961-8146-fbd28735e7af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 951.874100] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:e6:41:ee', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2624812a-9f9c-461d-8b5f-79bea90c7ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'a98dff4a-a9af-4961-8146-fbd28735e7af', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 951.884298] env[62753]: DEBUG oslo.service.loopingcall [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 951.884298] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 951.884803] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-38faceca-d4c7-4ad3-8c66-c782c2266758 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.906464] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 951.906464] env[62753]: value = "task-1332459" [ 951.906464] env[62753]: _type = "Task" [ 951.906464] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 951.915321] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332459, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 951.918652] env[62753]: DEBUG nova.compute.manager [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 951.949504] env[62753]: DEBUG nova.virt.hardware [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='f23b35a9bb79bdbcc40957c6dc697f5f',container_format='bare',created_at=2024-10-21T20:32:28Z,direct_url=,disk_format='vmdk',id=673e56af-bdea-458e-b52c-3abfd1149b15,min_disk=1,min_ram=0,name='tempest-test-snap-381536991',owner='e376836411ae4854965adf2923eaaffd',properties=ImageMetaProps,protected=,size=21334016,status='active',tags=,updated_at=2024-10-21T20:32:43Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 951.949922] env[62753]: DEBUG nova.virt.hardware [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 951.950179] env[62753]: DEBUG nova.virt.hardware [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 951.950493] env[62753]: DEBUG nova.virt.hardware [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 951.950735] env[62753]: DEBUG nova.virt.hardware [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 951.950973] env[62753]: DEBUG nova.virt.hardware [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 951.951347] env[62753]: DEBUG nova.virt.hardware [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 951.951624] env[62753]: DEBUG nova.virt.hardware [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 951.951900] env[62753]: DEBUG nova.virt.hardware [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 951.952188] env[62753]: DEBUG nova.virt.hardware [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 951.952494] env[62753]: DEBUG nova.virt.hardware [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 951.964246] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Creating linked-clone VM from snapshot {{(pid=62753) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 951.965515] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42ec50b-cc50-4b44-bb03-b6889da7ef40 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.970191] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-aafe7473-b45f-4cfb-832f-70a11cccfe95 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.982066] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3731a121-2a79-449b-a75e-2a5e29b8ccf0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 951.987050] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 951.987050] env[62753]: value = "task-1332460" [ 951.987050] env[62753]: _type = "Task" [ 951.987050] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.006507] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332460, 'name': CloneVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.184039] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332458, 'name': ReconfigVM_Task, 'duration_secs': 0.972788} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.184600] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Reconfigured VM instance instance-00000051 to attach disk [datastore2] c3679c66-2763-4a04-8d14-e103bd952798/c3679c66-2763-4a04-8d14-e103bd952798.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 952.185315] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57deacf3-a38f-4ab6-a1af-04d4cb9d1a5b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.195301] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 952.195301] env[62753]: value = "task-1332464" [ 952.195301] env[62753]: _type = "Task" [ 952.195301] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.208490] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332464, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.227057] env[62753]: DEBUG nova.objects.instance [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lazy-loading 'numa_topology' on Instance uuid 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 952.265128] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d9355eb5-d555-4058-bfe1-cec7b770eb29 tempest-ServersTestMultiNic-668169472 tempest-ServersTestMultiNic-668169472-project-member] Lock "f8470f38-0150-41c6-a3ee-aabb5e68d98b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.200s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 952.291779] env[62753]: DEBUG nova.compute.manager [req-607dd1bc-8094-4222-8f5c-ff7eeecee88b req-0113d845-9a77-45c4-9e1f-b0f39d436b5e service nova] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Received event network-changed-a98dff4a-a9af-4961-8146-fbd28735e7af {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 952.291984] env[62753]: DEBUG nova.compute.manager [req-607dd1bc-8094-4222-8f5c-ff7eeecee88b req-0113d845-9a77-45c4-9e1f-b0f39d436b5e service nova] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Refreshing instance network info cache due to event network-changed-a98dff4a-a9af-4961-8146-fbd28735e7af. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 952.292239] env[62753]: DEBUG oslo_concurrency.lockutils [req-607dd1bc-8094-4222-8f5c-ff7eeecee88b req-0113d845-9a77-45c4-9e1f-b0f39d436b5e service nova] Acquiring lock "refresh_cache-abd5ec6e-7b12-433b-9009-f9108ab4e318" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.292384] env[62753]: DEBUG oslo_concurrency.lockutils [req-607dd1bc-8094-4222-8f5c-ff7eeecee88b req-0113d845-9a77-45c4-9e1f-b0f39d436b5e service nova] Acquired lock "refresh_cache-abd5ec6e-7b12-433b-9009-f9108ab4e318" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.292552] env[62753]: DEBUG nova.network.neutron [req-607dd1bc-8094-4222-8f5c-ff7eeecee88b req-0113d845-9a77-45c4-9e1f-b0f39d436b5e service nova] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Refreshing network info cache for port a98dff4a-a9af-4961-8146-fbd28735e7af {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 952.418226] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332459, 'name': CreateVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.432054] env[62753]: DEBUG nova.network.neutron [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Successfully updated port: 2003ddeb-12c2-4450-a45b-cc180b041d34 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 952.500694] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332460, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.706803] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332464, 'name': Rename_Task, 'duration_secs': 0.193706} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.707162] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 952.707436] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5b9ed034-1da2-4712-8c53-948bfb77b134 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.716104] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 952.716104] env[62753]: value = "task-1332465" [ 952.716104] env[62753]: _type = "Task" [ 952.716104] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.725885] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332465, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.729453] env[62753]: INFO nova.compute.claims [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 952.920204] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332459, 'name': CreateVM_Task, 'duration_secs': 0.597327} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 952.922928] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 952.927355] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.927535] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.927912] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 952.928554] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e53161ac-d8f9-4a5d-876e-6535de2688a7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 952.935505] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "refresh_cache-ff4b295e-2eb2-42dd-b353-08df755f04ea" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 952.935708] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired lock "refresh_cache-ff4b295e-2eb2-42dd-b353-08df755f04ea" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 952.935884] env[62753]: DEBUG nova.network.neutron [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 952.937354] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 952.937354] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525b2648-c1aa-e122-4d0f-d8f68755431c" [ 952.937354] env[62753]: _type = "Task" [ 952.937354] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 952.948037] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525b2648-c1aa-e122-4d0f-d8f68755431c, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 952.999051] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332460, 'name': CloneVM_Task} progress is 95%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.141530] env[62753]: DEBUG nova.network.neutron [req-607dd1bc-8094-4222-8f5c-ff7eeecee88b req-0113d845-9a77-45c4-9e1f-b0f39d436b5e service nova] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Updated VIF entry in instance network info cache for port a98dff4a-a9af-4961-8146-fbd28735e7af. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 953.141919] env[62753]: DEBUG nova.network.neutron [req-607dd1bc-8094-4222-8f5c-ff7eeecee88b req-0113d845-9a77-45c4-9e1f-b0f39d436b5e service nova] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Updating instance_info_cache with network_info: [{"id": "a98dff4a-a9af-4961-8146-fbd28735e7af", "address": "fa:16:3e:e6:41:ee", "network": {"id": "35ed0ec0-821f-45b6-bc2d-015626bfc0a2", "bridge": "br-int", "label": "tempest-ServersTestJSON-444726172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5636da907ab343f9b42dbbd903d32283", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapa98dff4a-a9", "ovs_interfaceid": "a98dff4a-a9af-4961-8146-fbd28735e7af", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.227024] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332465, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.281603] env[62753]: DEBUG oslo_concurrency.lockutils [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "a401909c-5978-492b-a4fa-b17d38a5af44" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 953.281895] env[62753]: DEBUG oslo_concurrency.lockutils [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "a401909c-5978-492b-a4fa-b17d38a5af44" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 953.282136] env[62753]: DEBUG nova.compute.manager [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Going to confirm migration 1 {{(pid=62753) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 953.450144] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525b2648-c1aa-e122-4d0f-d8f68755431c, 'name': SearchDatastore_Task, 'duration_secs': 0.012349} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.450518] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.450886] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 953.451125] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.451357] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.451645] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 953.453099] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4146a8b1-e577-4b46-a1bd-103790f96c96 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.461564] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 953.461782] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 953.462563] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-66e8fa5d-c139-4998-a32d-85edf8a255da {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.468653] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 953.468653] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d17ddc-5f82-d659-eef9-5800a301ad8a" [ 953.468653] env[62753]: _type = "Task" [ 953.468653] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 953.475360] env[62753]: DEBUG nova.network.neutron [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 953.480423] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d17ddc-5f82-d659-eef9-5800a301ad8a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 953.499243] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332460, 'name': CloneVM_Task, 'duration_secs': 1.177637} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.499392] env[62753]: INFO nova.virt.vmwareapi.vmops [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Created linked-clone VM from snapshot [ 953.500078] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56594a1d-8f4a-435c-ae3f-5c54ce46be7d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.507994] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Uploading image 03c3734b-05a6-4c37-9f4e-406198dd6413 {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 953.532021] env[62753]: DEBUG oslo_vmware.rw_handles [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 953.532021] env[62753]: value = "vm-284634" [ 953.532021] env[62753]: _type = "VirtualMachine" [ 953.532021] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 953.532021] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1139e9c0-cdd2-4458-a2f7-961476bcfd6c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.538758] env[62753]: DEBUG oslo_vmware.rw_handles [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lease: (returnval){ [ 953.538758] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d7cf33-0e7e-d745-45fd-a917c0b8c6c1" [ 953.538758] env[62753]: _type = "HttpNfcLease" [ 953.538758] env[62753]: } obtained for exporting VM: (result){ [ 953.538758] env[62753]: value = "vm-284634" [ 953.538758] env[62753]: _type = "VirtualMachine" [ 953.538758] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 953.539046] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the lease: (returnval){ [ 953.539046] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d7cf33-0e7e-d745-45fd-a917c0b8c6c1" [ 953.539046] env[62753]: _type = "HttpNfcLease" [ 953.539046] env[62753]: } to be ready. {{(pid=62753) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 953.553215] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 953.553215] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d7cf33-0e7e-d745-45fd-a917c0b8c6c1" [ 953.553215] env[62753]: _type = "HttpNfcLease" [ 953.553215] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 953.621102] env[62753]: DEBUG nova.network.neutron [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Updating instance_info_cache with network_info: [{"id": "2003ddeb-12c2-4450-a45b-cc180b041d34", "address": "fa:16:3e:04:c8:f8", "network": {"id": "8c5edc3b-7faf-47f2-9736-7093f9ff72de", "bridge": "br-int", "label": "tempest-ImagesTestJSON-588052284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e376836411ae4854965adf2923eaaffd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76f377cd-5966-49b4-9210-907f592c694e", "external-id": "nsx-vlan-transportzone-124", "segmentation_id": 124, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2003ddeb-12", "ovs_interfaceid": "2003ddeb-12c2-4450-a45b-cc180b041d34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 953.645712] env[62753]: DEBUG oslo_concurrency.lockutils [req-607dd1bc-8094-4222-8f5c-ff7eeecee88b req-0113d845-9a77-45c4-9e1f-b0f39d436b5e service nova] Releasing lock "refresh_cache-abd5ec6e-7b12-433b-9009-f9108ab4e318" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 953.727991] env[62753]: DEBUG oslo_vmware.api [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332465, 'name': PowerOnVM_Task, 'duration_secs': 0.571006} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.728315] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 953.728563] env[62753]: INFO nova.compute.manager [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Took 8.75 seconds to spawn the instance on the hypervisor. [ 953.728760] env[62753]: DEBUG nova.compute.manager [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 953.729608] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-856f83c3-c8f9-48a2-b615-7c4aa40ccafd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.916296] env[62753]: DEBUG oslo_concurrency.lockutils [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "refresh_cache-a401909c-5978-492b-a4fa-b17d38a5af44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 953.916296] env[62753]: DEBUG oslo_concurrency.lockutils [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "refresh_cache-a401909c-5978-492b-a4fa-b17d38a5af44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 953.916296] env[62753]: DEBUG nova.network.neutron [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 953.916296] env[62753]: DEBUG nova.objects.instance [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lazy-loading 'info_cache' on Instance uuid a401909c-5978-492b-a4fa-b17d38a5af44 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 953.967821] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-802bc237-42b2-4967-82ee-e1dcdad755c3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.986129] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d17ddc-5f82-d659-eef9-5800a301ad8a, 'name': SearchDatastore_Task, 'duration_secs': 0.010556} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 953.987975] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f29f88e1-e859-4b6b-8ff9-1b05227ae537 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.991578] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6d869708-2bd7-439c-8a63-a0461686a5a1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 953.998366] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 953.998366] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52683924-8c0a-1414-0f9b-212e222a371b" [ 953.998366] env[62753]: _type = "Task" [ 953.998366] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.029055] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5d4bd11-d4ea-4c73-85dd-46b0c061abde {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.041861] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52683924-8c0a-1414-0f9b-212e222a371b, 'name': SearchDatastore_Task, 'duration_secs': 0.0142} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 954.045054] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.045335] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] abd5ec6e-7b12-433b-9009-f9108ab4e318/abd5ec6e-7b12-433b-9009-f9108ab4e318.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 954.046551] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc909b46-fe54-4953-a770-e3d69dfa505b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.051105] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-4308e2ce-4dfe-44a6-9a12-d6b285460db5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.056978] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 954.056978] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d7cf33-0e7e-d745-45fd-a917c0b8c6c1" [ 954.056978] env[62753]: _type = "HttpNfcLease" [ 954.056978] env[62753]: } is ready. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 954.065393] env[62753]: DEBUG oslo_vmware.rw_handles [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 954.065393] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d7cf33-0e7e-d745-45fd-a917c0b8c6c1" [ 954.065393] env[62753]: _type = "HttpNfcLease" [ 954.065393] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 954.065921] env[62753]: DEBUG nova.compute.provider_tree [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 954.068941] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a767c71-d78e-4bec-88c4-e62a63d57025 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.071894] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 954.071894] env[62753]: value = "task-1332467" [ 954.071894] env[62753]: _type = "Task" [ 954.071894] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.079798] env[62753]: DEBUG oslo_vmware.rw_handles [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527203da-b726-a702-98cc-f84eb8d56997/disk-0.vmdk from lease info. {{(pid=62753) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 954.079976] env[62753]: DEBUG oslo_vmware.rw_handles [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527203da-b726-a702-98cc-f84eb8d56997/disk-0.vmdk for reading. {{(pid=62753) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 954.142363] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Releasing lock "refresh_cache-ff4b295e-2eb2-42dd-b353-08df755f04ea" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 954.142708] env[62753]: DEBUG nova.compute.manager [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Instance network_info: |[{"id": "2003ddeb-12c2-4450-a45b-cc180b041d34", "address": "fa:16:3e:04:c8:f8", "network": {"id": "8c5edc3b-7faf-47f2-9736-7093f9ff72de", "bridge": "br-int", "label": "tempest-ImagesTestJSON-588052284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e376836411ae4854965adf2923eaaffd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76f377cd-5966-49b4-9210-907f592c694e", "external-id": "nsx-vlan-transportzone-124", "segmentation_id": 124, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2003ddeb-12", "ovs_interfaceid": "2003ddeb-12c2-4450-a45b-cc180b041d34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 954.149129] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:04:c8:f8', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '76f377cd-5966-49b4-9210-907f592c694e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2003ddeb-12c2-4450-a45b-cc180b041d34', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 954.154014] env[62753]: DEBUG oslo.service.loopingcall [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 954.154273] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332467, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.156069] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 954.156315] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-fdc2822d-9b01-42d6-a613-ad96c4c0d2ed {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.177350] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 954.177350] env[62753]: value = "task-1332469" [ 954.177350] env[62753]: _type = "Task" [ 954.177350] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 954.207479] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-27abf3de-c610-4d6b-8b6d-08c41d59f84f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 954.208217] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332469, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.258155] env[62753]: INFO nova.compute.manager [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Took 19.63 seconds to build instance. [ 954.518032] env[62753]: DEBUG nova.compute.manager [req-04a6bf98-f1c3-4792-9a98-5c57fa5b2e0c req-0d571785-641b-4dd4-80e1-3d55888cbb64 service nova] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Received event network-vif-plugged-2003ddeb-12c2-4450-a45b-cc180b041d34 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.518301] env[62753]: DEBUG oslo_concurrency.lockutils [req-04a6bf98-f1c3-4792-9a98-5c57fa5b2e0c req-0d571785-641b-4dd4-80e1-3d55888cbb64 service nova] Acquiring lock "ff4b295e-2eb2-42dd-b353-08df755f04ea-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 954.518510] env[62753]: DEBUG oslo_concurrency.lockutils [req-04a6bf98-f1c3-4792-9a98-5c57fa5b2e0c req-0d571785-641b-4dd4-80e1-3d55888cbb64 service nova] Lock "ff4b295e-2eb2-42dd-b353-08df755f04ea-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 954.518710] env[62753]: DEBUG oslo_concurrency.lockutils [req-04a6bf98-f1c3-4792-9a98-5c57fa5b2e0c req-0d571785-641b-4dd4-80e1-3d55888cbb64 service nova] Lock "ff4b295e-2eb2-42dd-b353-08df755f04ea-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 954.518854] env[62753]: DEBUG nova.compute.manager [req-04a6bf98-f1c3-4792-9a98-5c57fa5b2e0c req-0d571785-641b-4dd4-80e1-3d55888cbb64 service nova] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] No waiting events found dispatching network-vif-plugged-2003ddeb-12c2-4450-a45b-cc180b041d34 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 954.519350] env[62753]: WARNING nova.compute.manager [req-04a6bf98-f1c3-4792-9a98-5c57fa5b2e0c req-0d571785-641b-4dd4-80e1-3d55888cbb64 service nova] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Received unexpected event network-vif-plugged-2003ddeb-12c2-4450-a45b-cc180b041d34 for instance with vm_state building and task_state spawning. [ 954.519587] env[62753]: DEBUG nova.compute.manager [req-04a6bf98-f1c3-4792-9a98-5c57fa5b2e0c req-0d571785-641b-4dd4-80e1-3d55888cbb64 service nova] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Received event network-changed-2003ddeb-12c2-4450-a45b-cc180b041d34 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 954.519792] env[62753]: DEBUG nova.compute.manager [req-04a6bf98-f1c3-4792-9a98-5c57fa5b2e0c req-0d571785-641b-4dd4-80e1-3d55888cbb64 service nova] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Refreshing instance network info cache due to event network-changed-2003ddeb-12c2-4450-a45b-cc180b041d34. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 954.520134] env[62753]: DEBUG oslo_concurrency.lockutils [req-04a6bf98-f1c3-4792-9a98-5c57fa5b2e0c req-0d571785-641b-4dd4-80e1-3d55888cbb64 service nova] Acquiring lock "refresh_cache-ff4b295e-2eb2-42dd-b353-08df755f04ea" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 954.520342] env[62753]: DEBUG oslo_concurrency.lockutils [req-04a6bf98-f1c3-4792-9a98-5c57fa5b2e0c req-0d571785-641b-4dd4-80e1-3d55888cbb64 service nova] Acquired lock "refresh_cache-ff4b295e-2eb2-42dd-b353-08df755f04ea" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 954.520546] env[62753]: DEBUG nova.network.neutron [req-04a6bf98-f1c3-4792-9a98-5c57fa5b2e0c req-0d571785-641b-4dd4-80e1-3d55888cbb64 service nova] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Refreshing network info cache for port 2003ddeb-12c2-4450-a45b-cc180b041d34 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 954.574187] env[62753]: DEBUG nova.scheduler.client.report [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 954.590502] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332467, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.689590] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332469, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 954.758458] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e8bd17b0-ebe3-4ebb-aa47-ef06aab0827c tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "c3679c66-2763-4a04-8d14-e103bd952798" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 21.148s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.087999] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.366s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.091101] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 7.578s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.091396] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 955.091661] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62753) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 955.093756] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31ae5da9-a2b2-44f7-a36a-719c8f62e9ee {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.104034] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332467, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.589278} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.105728] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] abd5ec6e-7b12-433b-9009-f9108ab4e318/abd5ec6e-7b12-433b-9009-f9108ab4e318.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 955.106588] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 955.107041] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-919095eb-2672-4e07-84e0-14be8679a214 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.120564] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-781cc338-4af4-4fd9-869d-0c4e391639a2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.128030] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 955.128030] env[62753]: value = "task-1332470" [ 955.128030] env[62753]: _type = "Task" [ 955.128030] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.148607] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61ebd0c4-64f8-4d4a-a4da-a07b3d33ef62 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.158499] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332470, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.167182] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a53a3557-4546-49c1-9ef0-c611df110089 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.209025] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180320MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=62753) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 955.209351] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 955.209607] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 955.215379] env[62753]: INFO nova.network.neutron [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Updating port 7fb046f4-e470-4f6d-8fa3-73f3b0157020 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 955.231797] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332469, 'name': CreateVM_Task, 'duration_secs': 0.731013} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.232055] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 955.232984] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/673e56af-bdea-458e-b52c-3abfd1149b15" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.233254] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired lock "[datastore2] devstack-image-cache_base/673e56af-bdea-458e-b52c-3abfd1149b15" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.233814] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/673e56af-bdea-458e-b52c-3abfd1149b15" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 955.234162] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b4dc033-0d23-41ba-9709-64504b407884 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.241469] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 955.241469] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c961e3-f7e2-75d4-01f7-59d03f94c4a6" [ 955.241469] env[62753]: _type = "Task" [ 955.241469] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.252902] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c961e3-f7e2-75d4-01f7-59d03f94c4a6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.378402] env[62753]: DEBUG nova.network.neutron [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Updating instance_info_cache with network_info: [{"id": "7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36", "address": "fa:16:3e:52:63:e3", "network": {"id": "2823647f-d32d-4a1c-9e02-7577016a260e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1594498922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6c8a6bc946a469fa85cc3ab80d7333a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7df5849e-e6", "ovs_interfaceid": "7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.641099] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332470, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.08906} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 955.641780] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 955.643578] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e76975f2-e0a6-4876-8008-2d466de82f84 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.677316] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Reconfiguring VM instance instance-00000052 to attach disk [datastore2] abd5ec6e-7b12-433b-9009-f9108ab4e318/abd5ec6e-7b12-433b-9009-f9108ab4e318.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 955.677906] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0df6f461-6dad-451f-bae5-79bee44c11fb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.700946] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 955.700946] env[62753]: value = "task-1332471" [ 955.700946] env[62753]: _type = "Task" [ 955.700946] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.711911] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332471, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.755260] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Releasing lock "[datastore2] devstack-image-cache_base/673e56af-bdea-458e-b52c-3abfd1149b15" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.755619] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Processing image 673e56af-bdea-458e-b52c-3abfd1149b15 {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 955.755930] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/673e56af-bdea-458e-b52c-3abfd1149b15/673e56af-bdea-458e-b52c-3abfd1149b15.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 955.759402] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired lock "[datastore2] devstack-image-cache_base/673e56af-bdea-458e-b52c-3abfd1149b15/673e56af-bdea-458e-b52c-3abfd1149b15.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 955.759402] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 955.760480] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-80428f77-daf3-4980-909f-95e57eeed188 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.772057] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 955.772384] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 955.773888] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-91350956-b98b-4ee4-87b7-bb65cadfa7cb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 955.778588] env[62753]: DEBUG nova.network.neutron [req-04a6bf98-f1c3-4792-9a98-5c57fa5b2e0c req-0d571785-641b-4dd4-80e1-3d55888cbb64 service nova] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Updated VIF entry in instance network info cache for port 2003ddeb-12c2-4450-a45b-cc180b041d34. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 955.779074] env[62753]: DEBUG nova.network.neutron [req-04a6bf98-f1c3-4792-9a98-5c57fa5b2e0c req-0d571785-641b-4dd4-80e1-3d55888cbb64 service nova] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Updating instance_info_cache with network_info: [{"id": "2003ddeb-12c2-4450-a45b-cc180b041d34", "address": "fa:16:3e:04:c8:f8", "network": {"id": "8c5edc3b-7faf-47f2-9736-7093f9ff72de", "bridge": "br-int", "label": "tempest-ImagesTestJSON-588052284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e376836411ae4854965adf2923eaaffd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76f377cd-5966-49b4-9210-907f592c694e", "external-id": "nsx-vlan-transportzone-124", "segmentation_id": 124, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2003ddeb-12", "ovs_interfaceid": "2003ddeb-12c2-4450-a45b-cc180b041d34", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 955.787501] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 955.787501] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52cfa646-58d7-aec7-2f2b-793b8b372e7a" [ 955.787501] env[62753]: _type = "Task" [ 955.787501] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 955.802495] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52cfa646-58d7-aec7-2f2b-793b8b372e7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 955.882364] env[62753]: DEBUG oslo_concurrency.lockutils [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "refresh_cache-a401909c-5978-492b-a4fa-b17d38a5af44" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 955.882924] env[62753]: DEBUG nova.objects.instance [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lazy-loading 'migration_context' on Instance uuid a401909c-5978-492b-a4fa-b17d38a5af44 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 956.216530] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332471, 'name': ReconfigVM_Task, 'duration_secs': 0.48645} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.216954] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Reconfigured VM instance instance-00000052 to attach disk [datastore2] abd5ec6e-7b12-433b-9009-f9108ab4e318/abd5ec6e-7b12-433b-9009-f9108ab4e318.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 956.217708] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-6e5be006-3a6d-4dc3-92cf-14897c2c6cd7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.226758] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 956.226758] env[62753]: value = "task-1332472" [ 956.226758] env[62753]: _type = "Task" [ 956.226758] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.232436] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Applying migration context for instance a401909c-5978-492b-a4fa-b17d38a5af44 as it has an incoming, in-progress migration 3c736453-0224-4471-9d77-b23fddf9618f. Migration status is confirming {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 956.234314] env[62753]: INFO nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Updating resource usage from migration 3c736453-0224-4471-9d77-b23fddf9618f [ 956.241014] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332472, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.264432] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 4d20e083-2959-453a-8875-47955bc02613 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 956.264889] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance c73b1ae0-dc98-47f7-babf-e96169384785 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 956.264889] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 77cdd901-cc96-4a6f-a696-a65f54a96b1a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 956.264889] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 956.265083] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance b5674964-1928-4ecb-b1fd-8f60a94b4270 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 956.265132] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 8066aeb2-7be4-46e3-aba6-124fef342a49 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 956.265250] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 628239e8-a447-4183-bed9-5550e53fda2b actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 956.265364] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Migration 3c736453-0224-4471-9d77-b23fddf9618f is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 956.265479] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance a401909c-5978-492b-a4fa-b17d38a5af44 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 956.265592] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance c3679c66-2763-4a04-8d14-e103bd952798 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 956.265706] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 956.265844] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance abd5ec6e-7b12-433b-9009-f9108ab4e318 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 956.265965] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance ff4b295e-2eb2-42dd-b353-08df755f04ea actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 956.266393] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Total usable vcpus: 48, total allocated vcpus: 13 {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 956.266576] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=3072MB phys_disk=200GB used_disk=13GB total_vcpus=48 used_vcpus=13 pci_stats=[] {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 956.284688] env[62753]: DEBUG oslo_concurrency.lockutils [req-04a6bf98-f1c3-4792-9a98-5c57fa5b2e0c req-0d571785-641b-4dd4-80e1-3d55888cbb64 service nova] Releasing lock "refresh_cache-ff4b295e-2eb2-42dd-b353-08df755f04ea" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 956.310468] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Preparing fetch location {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 956.310895] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Fetch image to [datastore2] OSTACK_IMG_23afe077-5e3b-4381-907e-66033b929d0e/OSTACK_IMG_23afe077-5e3b-4381-907e-66033b929d0e.vmdk {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 956.310957] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Downloading stream optimized image 673e56af-bdea-458e-b52c-3abfd1149b15 to [datastore2] OSTACK_IMG_23afe077-5e3b-4381-907e-66033b929d0e/OSTACK_IMG_23afe077-5e3b-4381-907e-66033b929d0e.vmdk on the data store datastore2 as vApp {{(pid=62753) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 956.311172] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Downloading image file data 673e56af-bdea-458e-b52c-3abfd1149b15 to the ESX as VM named 'OSTACK_IMG_23afe077-5e3b-4381-907e-66033b929d0e' {{(pid=62753) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 956.388255] env[62753]: DEBUG nova.objects.base [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 956.389779] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9a3ffcb9-bddf-4a91-a8f2-4df4f63d3b0f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.415612] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d3ad9c66-4394-4995-b919-2c62b12b44be {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.417603] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Volume attach. Driver type: vmdk {{(pid=62753) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 956.417835] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284636', 'volume_id': '368def30-47d9-476e-8669-d2beafd3fba1', 'name': 'volume-368def30-47d9-476e-8669-d2beafd3fba1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '77cdd901-cc96-4a6f-a696-a65f54a96b1a', 'attached_at': '', 'detached_at': '', 'volume_id': '368def30-47d9-476e-8669-d2beafd3fba1', 'serial': '368def30-47d9-476e-8669-d2beafd3fba1'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 956.418900] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc39fc8f-6148-4ff7-b24f-4d5a19034986 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.427150] env[62753]: DEBUG oslo_vmware.api [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 956.427150] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]521133f0-f266-e68d-1c01-87c823a61cda" [ 956.427150] env[62753]: _type = "Task" [ 956.427150] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.446121] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5e251d5-5d5b-451f-a023-81ba30457d3f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.456350] env[62753]: DEBUG oslo_vmware.api [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]521133f0-f266-e68d-1c01-87c823a61cda, 'name': SearchDatastore_Task, 'duration_secs': 0.021078} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.478023] env[62753]: DEBUG oslo_concurrency.lockutils [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 956.486635] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Reconfiguring VM instance instance-00000049 to attach disk [datastore2] volume-368def30-47d9-476e-8669-d2beafd3fba1/volume-368def30-47d9-476e-8669-d2beafd3fba1.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 956.490745] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1a0f7b92-7591-4112-85d3-3c23890b06f6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.511494] env[62753]: DEBUG oslo_vmware.api [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 956.511494] env[62753]: value = "task-1332473" [ 956.511494] env[62753]: _type = "Task" [ 956.511494] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.524338] env[62753]: DEBUG oslo_vmware.api [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332473, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 956.577074] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d98784ee-04be-4dd3-8e45-d58aea86528f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.586924] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9414c09a-3c83-4030-a58c-1df6ddd75474 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.619268] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a73f6fd-b6bb-41c5-a663-d50c290d8051 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.628182] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b56d715-9ca8-4bef-9b0c-fdf9b741662b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.648670] env[62753]: DEBUG nova.compute.provider_tree [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 956.740946] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332472, 'name': Rename_Task, 'duration_secs': 0.207906} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 956.741424] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 956.741769] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4125b801-fc14-4d57-aa4d-0c05ff4e9734 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 956.751478] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 956.751478] env[62753]: value = "task-1332474" [ 956.751478] env[62753]: _type = "Task" [ 956.751478] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 956.761936] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332474, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.027819] env[62753]: DEBUG oslo_vmware.api [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332473, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.088624] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 957.088675] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquired lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 957.088920] env[62753]: DEBUG nova.network.neutron [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 957.152805] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 957.262382] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332474, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.524021] env[62753]: DEBUG oslo_vmware.api [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332473, 'name': ReconfigVM_Task, 'duration_secs': 0.632986} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.524021] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Reconfigured VM instance instance-00000049 to attach disk [datastore2] volume-368def30-47d9-476e-8669-d2beafd3fba1/volume-368def30-47d9-476e-8669-d2beafd3fba1.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 957.528821] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-60364d09-38df-4fb6-9ca7-b9df16e7de81 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.544799] env[62753]: DEBUG oslo_vmware.api [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 957.544799] env[62753]: value = "task-1332475" [ 957.544799] env[62753]: _type = "Task" [ 957.544799] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 957.554234] env[62753]: DEBUG oslo_vmware.api [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332475, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 957.658257] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 957.658485] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.449s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 957.658818] env[62753]: DEBUG oslo_concurrency.lockutils [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 1.181s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 957.765183] env[62753]: DEBUG oslo_vmware.api [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332474, 'name': PowerOnVM_Task, 'duration_secs': 0.676913} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 957.766649] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 957.766649] env[62753]: INFO nova.compute.manager [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Took 8.17 seconds to spawn the instance on the hypervisor. [ 957.766649] env[62753]: DEBUG nova.compute.manager [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 957.767377] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d00ba8ae-8ff6-4600-a165-873ceac4a06e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 957.841731] env[62753]: DEBUG nova.network.neutron [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Updating instance_info_cache with network_info: [{"id": "7fb046f4-e470-4f6d-8fa3-73f3b0157020", "address": "fa:16:3e:6f:26:2c", "network": {"id": "fc646f7f-70ba-4671-8476-3b0ece1e9041", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-853174775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d0eb0b3ba504f59a85cb021da2b47b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fb046f4-e4", "ovs_interfaceid": "7fb046f4-e470-4f6d-8fa3-73f3b0157020", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 958.056673] env[62753]: DEBUG oslo_vmware.api [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332475, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.295911] env[62753]: INFO nova.compute.manager [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Took 18.45 seconds to build instance. [ 958.344722] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Releasing lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 958.360298] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94344d3b-ddc1-49c2-9496-01d21e70b8c2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.369925] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-20903b48-749c-4300-8394-7e25dbc8ca81 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.405544] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fc0e7870-ea00-450e-bab1-e4f42e919930 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.414102] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13bfe77f-ecc6-4336-90cf-9e520813b697 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.428495] env[62753]: DEBUG nova.compute.provider_tree [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 958.557333] env[62753]: DEBUG oslo_vmware.api [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332475, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 958.798309] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3689434c-3c30-4ef7-8715-7a4205481e74 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "abd5ec6e-7b12-433b-9009-f9108ab4e318" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.962s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 958.845334] env[62753]: DEBUG nova.compute.manager [req-90384d50-12e7-4190-bd3f-fe0fc73deb40 req-fb4e2d43-ccf4-404c-9bc5-0bb0e63ccc0f service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Received event network-changed-12ca481a-0cb3-465e-b04c-802ae3ce6438 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 958.845334] env[62753]: DEBUG nova.compute.manager [req-90384d50-12e7-4190-bd3f-fe0fc73deb40 req-fb4e2d43-ccf4-404c-9bc5-0bb0e63ccc0f service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Refreshing instance network info cache due to event network-changed-12ca481a-0cb3-465e-b04c-802ae3ce6438. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 958.845665] env[62753]: DEBUG oslo_concurrency.lockutils [req-90384d50-12e7-4190-bd3f-fe0fc73deb40 req-fb4e2d43-ccf4-404c-9bc5-0bb0e63ccc0f service nova] Acquiring lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 958.845927] env[62753]: DEBUG oslo_concurrency.lockutils [req-90384d50-12e7-4190-bd3f-fe0fc73deb40 req-fb4e2d43-ccf4-404c-9bc5-0bb0e63ccc0f service nova] Acquired lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 958.846189] env[62753]: DEBUG nova.network.neutron [req-90384d50-12e7-4190-bd3f-fe0fc73deb40 req-fb4e2d43-ccf4-404c-9bc5-0bb0e63ccc0f service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Refreshing network info cache for port 12ca481a-0cb3-465e-b04c-802ae3ce6438 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 958.855662] env[62753]: DEBUG nova.virt.hardware [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='e674fce39ed22316b3beb69e9070a445',container_format='bare',created_at=2024-10-21T20:32:16Z,direct_url=,disk_format='vmdk',id=061f0060-350a-4bf7-a725-58acba59de2b,min_disk=1,min_ram=0,name='tempest-AttachVolumeShelveTestJSON-server-933654694-shelved',owner='4d0eb0b3ba504f59a85cb021da2b47b2',properties=ImageMetaProps,protected=,size=31667712,status='active',tags=,updated_at=2024-10-21T20:32:33Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 958.855960] env[62753]: DEBUG nova.virt.hardware [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 958.856235] env[62753]: DEBUG nova.virt.hardware [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 958.856492] env[62753]: DEBUG nova.virt.hardware [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 958.857140] env[62753]: DEBUG nova.virt.hardware [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 958.857140] env[62753]: DEBUG nova.virt.hardware [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 958.857257] env[62753]: DEBUG nova.virt.hardware [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 958.857469] env[62753]: DEBUG nova.virt.hardware [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 958.857754] env[62753]: DEBUG nova.virt.hardware [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 958.858039] env[62753]: DEBUG nova.virt.hardware [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 958.858326] env[62753]: DEBUG nova.virt.hardware [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 958.863282] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5b6ad81-f57d-4de7-98fa-6489a0d1666d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.875082] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00d143b0-5782-4d1a-b7bc-2f8f5846acc9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.881998] env[62753]: DEBUG oslo_vmware.rw_handles [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 958.881998] env[62753]: value = "resgroup-9" [ 958.881998] env[62753]: _type = "ResourcePool" [ 958.881998] env[62753]: }. {{(pid=62753) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 958.883364] env[62753]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-9d8c7263-c7fd-404c-8796-466b5b9aaa75 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.919436] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6f:26:2c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '255460d5-71d4-4bfd-87f1-acc10085db7f', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7fb046f4-e470-4f6d-8fa3-73f3b0157020', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 958.927556] env[62753]: DEBUG oslo.service.loopingcall [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 958.929647] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 958.930073] env[62753]: DEBUG oslo_vmware.rw_handles [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lease: (returnval){ [ 958.930073] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f9f510-baeb-c350-ed39-a1d4e81ebffb" [ 958.930073] env[62753]: _type = "HttpNfcLease" [ 958.930073] env[62753]: } obtained for vApp import into resource pool (val){ [ 958.930073] env[62753]: value = "resgroup-9" [ 958.930073] env[62753]: _type = "ResourcePool" [ 958.930073] env[62753]: }. {{(pid=62753) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 958.930409] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the lease: (returnval){ [ 958.930409] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f9f510-baeb-c350-ed39-a1d4e81ebffb" [ 958.930409] env[62753]: _type = "HttpNfcLease" [ 958.930409] env[62753]: } to be ready. {{(pid=62753) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 958.930611] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-757baac1-9df8-4c55-8cec-4b9f94342486 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 958.945381] env[62753]: DEBUG nova.scheduler.client.report [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 958.957333] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 958.957333] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f9f510-baeb-c350-ed39-a1d4e81ebffb" [ 958.957333] env[62753]: _type = "HttpNfcLease" [ 958.957333] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 958.958699] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 958.958699] env[62753]: value = "task-1332477" [ 958.958699] env[62753]: _type = "Task" [ 958.958699] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 958.967743] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332477, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.062330] env[62753]: DEBUG oslo_vmware.api [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332475, 'name': ReconfigVM_Task, 'duration_secs': 1.188485} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 959.062780] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284636', 'volume_id': '368def30-47d9-476e-8669-d2beafd3fba1', 'name': 'volume-368def30-47d9-476e-8669-d2beafd3fba1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '77cdd901-cc96-4a6f-a696-a65f54a96b1a', 'attached_at': '', 'detached_at': '', 'volume_id': '368def30-47d9-476e-8669-d2beafd3fba1', 'serial': '368def30-47d9-476e-8669-d2beafd3fba1'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 959.337891] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "4d20e083-2959-453a-8875-47955bc02613" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.338381] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "4d20e083-2959-453a-8875-47955bc02613" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.338675] env[62753]: INFO nova.compute.manager [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Shelving [ 959.464789] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 959.464789] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f9f510-baeb-c350-ed39-a1d4e81ebffb" [ 959.464789] env[62753]: _type = "HttpNfcLease" [ 959.464789] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 959.471199] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332477, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.572445] env[62753]: DEBUG nova.compute.manager [req-61c834cf-3082-4d14-b76d-5868865f4332 req-4f57eb53-5caf-461c-b5bd-ebe27727ebfc service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Received event network-vif-plugged-7fb046f4-e470-4f6d-8fa3-73f3b0157020 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 959.572445] env[62753]: DEBUG oslo_concurrency.lockutils [req-61c834cf-3082-4d14-b76d-5868865f4332 req-4f57eb53-5caf-461c-b5bd-ebe27727ebfc service nova] Acquiring lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 959.572445] env[62753]: DEBUG oslo_concurrency.lockutils [req-61c834cf-3082-4d14-b76d-5868865f4332 req-4f57eb53-5caf-461c-b5bd-ebe27727ebfc service nova] Lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 959.572753] env[62753]: DEBUG oslo_concurrency.lockutils [req-61c834cf-3082-4d14-b76d-5868865f4332 req-4f57eb53-5caf-461c-b5bd-ebe27727ebfc service nova] Lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.573562] env[62753]: DEBUG nova.compute.manager [req-61c834cf-3082-4d14-b76d-5868865f4332 req-4f57eb53-5caf-461c-b5bd-ebe27727ebfc service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] No waiting events found dispatching network-vif-plugged-7fb046f4-e470-4f6d-8fa3-73f3b0157020 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 959.574373] env[62753]: WARNING nova.compute.manager [req-61c834cf-3082-4d14-b76d-5868865f4332 req-4f57eb53-5caf-461c-b5bd-ebe27727ebfc service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Received unexpected event network-vif-plugged-7fb046f4-e470-4f6d-8fa3-73f3b0157020 for instance with vm_state shelved_offloaded and task_state spawning. [ 959.846737] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 959.847054] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-efd38dca-a957-41ac-aac7-15dbbae43e3f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.856473] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 959.856473] env[62753]: value = "task-1332478" [ 959.856473] env[62753]: _type = "Task" [ 959.856473] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 959.869442] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332478, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 959.935667] env[62753]: DEBUG nova.network.neutron [req-90384d50-12e7-4190-bd3f-fe0fc73deb40 req-fb4e2d43-ccf4-404c-9bc5-0bb0e63ccc0f service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Updated VIF entry in instance network info cache for port 12ca481a-0cb3-465e-b04c-802ae3ce6438. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 959.936308] env[62753]: DEBUG nova.network.neutron [req-90384d50-12e7-4190-bd3f-fe0fc73deb40 req-fb4e2d43-ccf4-404c-9bc5-0bb0e63ccc0f service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Updating instance_info_cache with network_info: [{"id": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "address": "fa:16:3e:57:5d:75", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12ca481a-0c", "ovs_interfaceid": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 959.962834] env[62753]: DEBUG oslo_concurrency.lockutils [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.303s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 959.965736] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 959.965736] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f9f510-baeb-c350-ed39-a1d4e81ebffb" [ 959.965736] env[62753]: _type = "HttpNfcLease" [ 959.965736] env[62753]: } is ready. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 959.969556] env[62753]: DEBUG oslo_vmware.rw_handles [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 959.969556] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f9f510-baeb-c350-ed39-a1d4e81ebffb" [ 959.969556] env[62753]: _type = "HttpNfcLease" [ 959.969556] env[62753]: }. {{(pid=62753) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 959.973976] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f2bcf1d-289e-4238-acd7-836eabc922df {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 959.982139] env[62753]: DEBUG oslo_vmware.rw_handles [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Found VMDK URL: https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e85ed3-0b7b-70a9-f231-32e43cafc227/disk-0.vmdk from lease info. {{(pid=62753) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 959.982349] env[62753]: DEBUG oslo_vmware.rw_handles [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Creating HTTP connection to write to file with size = 21334016 and URL = https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e85ed3-0b7b-70a9-f231-32e43cafc227/disk-0.vmdk. {{(pid=62753) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 959.986984] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332477, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.049548] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Acquiring lock "d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.049859] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Lock "d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.050200] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Acquiring lock "d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.050526] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Lock "d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.050779] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Lock "d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.054419] env[62753]: INFO nova.compute.manager [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Terminating instance [ 960.058617] env[62753]: DEBUG nova.compute.manager [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 960.059150] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 960.061248] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b871a032-f4b9-46c6-9c2f-89cde6c08547 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.069413] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-b48b78c2-214c-4f0b-a080-7a298041e062 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.075347] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 960.076713] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a8f8aef4-cecc-4150-aad8-1efe656929eb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.086027] env[62753]: DEBUG oslo_vmware.api [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Waiting for the task: (returnval){ [ 960.086027] env[62753]: value = "task-1332479" [ 960.086027] env[62753]: _type = "Task" [ 960.086027] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.096894] env[62753]: DEBUG oslo_vmware.api [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': task-1332479, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.124900] env[62753]: DEBUG nova.objects.instance [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lazy-loading 'flavor' on Instance uuid 77cdd901-cc96-4a6f-a696-a65f54a96b1a {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 960.194635] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 960.194953] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 960.370701] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332478, 'name': PowerOffVM_Task, 'duration_secs': 0.333896} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.374662] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 960.376148] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0bbdc3e0-6db2-4355-8fbe-330aeb29375a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.402421] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1921e020-e36e-45b4-a579-8ca1d3d50e3c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.441309] env[62753]: DEBUG oslo_vmware.rw_handles [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc73a9-2e4b-5918-e94e-09342a64ac74/disk-0.vmdk. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 960.442879] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab7eecdd-b2db-42eb-92d9-367b947b69ac {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.448980] env[62753]: DEBUG oslo_concurrency.lockutils [req-90384d50-12e7-4190-bd3f-fe0fc73deb40 req-fb4e2d43-ccf4-404c-9bc5-0bb0e63ccc0f service nova] Releasing lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.455649] env[62753]: DEBUG oslo_vmware.rw_handles [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc73a9-2e4b-5918-e94e-09342a64ac74/disk-0.vmdk is in state: ready. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 960.455853] env[62753]: ERROR oslo_vmware.rw_handles [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Aborting lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc73a9-2e4b-5918-e94e-09342a64ac74/disk-0.vmdk due to incomplete transfer. [ 960.457983] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-96932d49-7c71-422a-bfe6-75a128203202 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.479164] env[62753]: DEBUG oslo_vmware.rw_handles [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Closed VMDK read handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52cc73a9-2e4b-5918-e94e-09342a64ac74/disk-0.vmdk. {{(pid=62753) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 960.479444] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Uploaded image 1bf6f512-055c-454e-b20a-aa68af11c3df to the Glance image server {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 960.482590] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Destroying the VM {{(pid=62753) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 960.483859] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-4baac93e-c660-4d95-b3bf-eb78f3b3f5fa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.491804] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332477, 'name': CreateVM_Task, 'duration_secs': 1.190431} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.491954] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 960.493270] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/061f0060-350a-4bf7-a725-58acba59de2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.493270] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquired lock "[datastore1] devstack-image-cache_base/061f0060-350a-4bf7-a725-58acba59de2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.493270] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/061f0060-350a-4bf7-a725-58acba59de2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 960.495076] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fb3f7dbe-0284-411c-be82-e1757d27caf2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.499228] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 960.499228] env[62753]: value = "task-1332480" [ 960.499228] env[62753]: _type = "Task" [ 960.499228] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.510490] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 960.510490] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c051bf-f3a9-c491-e116-931ec52e9c0d" [ 960.510490] env[62753]: _type = "Task" [ 960.510490] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.523238] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332480, 'name': Destroy_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.532738] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Releasing lock "[datastore1] devstack-image-cache_base/061f0060-350a-4bf7-a725-58acba59de2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 960.533053] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Processing image 061f0060-350a-4bf7-a725-58acba59de2b {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 960.533280] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/061f0060-350a-4bf7-a725-58acba59de2b/061f0060-350a-4bf7-a725-58acba59de2b.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 960.533446] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquired lock "[datastore1] devstack-image-cache_base/061f0060-350a-4bf7-a725-58acba59de2b/061f0060-350a-4bf7-a725-58acba59de2b.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 960.533715] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 960.535659] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ca8fe1b7-8e2b-4b9c-a173-55266b4da26d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.548121] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 960.548431] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 960.552057] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7926f5a7-8e3a-4f71-9d20-7c9a904dec46 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.559380] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 960.559380] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52baea60-9f47-2f73-fb0e-7f1750ff4b45" [ 960.559380] env[62753]: _type = "Task" [ 960.559380] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.562418] env[62753]: INFO nova.scheduler.client.report [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Deleted allocation for migration 3c736453-0224-4471-9d77-b23fddf9618f [ 960.581245] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Preparing fetch location {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 960.581716] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Fetch image to [datastore1] OSTACK_IMG_715833fa-e566-42a9-8715-2235ec81fd4d/OSTACK_IMG_715833fa-e566-42a9-8715-2235ec81fd4d.vmdk {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 960.581920] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Downloading stream optimized image 061f0060-350a-4bf7-a725-58acba59de2b to [datastore1] OSTACK_IMG_715833fa-e566-42a9-8715-2235ec81fd4d/OSTACK_IMG_715833fa-e566-42a9-8715-2235ec81fd4d.vmdk on the data store datastore1 as vApp {{(pid=62753) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 960.582775] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Downloading image file data 061f0060-350a-4bf7-a725-58acba59de2b to the ESX as VM named 'OSTACK_IMG_715833fa-e566-42a9-8715-2235ec81fd4d' {{(pid=62753) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 960.600894] env[62753]: DEBUG oslo_vmware.api [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': task-1332479, 'name': PowerOffVM_Task, 'duration_secs': 0.237273} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 960.603426] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 960.604190] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 960.604190] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-5979db02-7d96-448a-b9e0-96dd842e91d2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.646207] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14704b53-b836-4c89-ad39-e9a0cc6ff250 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.847s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 960.661329] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 960.661688] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Starting heal instance info cache {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 960.663024] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Rebuilding the list of instances to heal {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 960.696680] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 960.696680] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 960.696680] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Deleting the datastore file [datastore2] d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 960.697997] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d6424309-4afe-4aec-bfe9-f4ec8cc08a69 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.702819] env[62753]: DEBUG nova.compute.manager [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 960.706625] env[62753]: DEBUG oslo_vmware.rw_handles [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 960.706625] env[62753]: value = "resgroup-9" [ 960.706625] env[62753]: _type = "ResourcePool" [ 960.706625] env[62753]: }. {{(pid=62753) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 960.710823] env[62753]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-efeb77fe-ba54-4fe6-871c-2d5267dcabef {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.741720] env[62753]: DEBUG oslo_vmware.api [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Waiting for the task: (returnval){ [ 960.741720] env[62753]: value = "task-1332482" [ 960.741720] env[62753]: _type = "Task" [ 960.741720] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.744342] env[62753]: DEBUG oslo_vmware.rw_handles [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lease: (returnval){ [ 960.744342] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c60a5d-4376-adca-3113-efef5b3f4c7b" [ 960.744342] env[62753]: _type = "HttpNfcLease" [ 960.744342] env[62753]: } obtained for vApp import into resource pool (val){ [ 960.744342] env[62753]: value = "resgroup-9" [ 960.744342] env[62753]: _type = "ResourcePool" [ 960.744342] env[62753]: }. {{(pid=62753) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 960.744813] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the lease: (returnval){ [ 960.744813] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c60a5d-4376-adca-3113-efef5b3f4c7b" [ 960.744813] env[62753]: _type = "HttpNfcLease" [ 960.744813] env[62753]: } to be ready. {{(pid=62753) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 960.759966] env[62753]: DEBUG oslo_vmware.api [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': task-1332482, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.761783] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 960.761783] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c60a5d-4376-adca-3113-efef5b3f4c7b" [ 960.761783] env[62753]: _type = "HttpNfcLease" [ 960.761783] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 960.916226] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Creating Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 960.917142] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-add0da78-2492-420c-8d52-0831d3d22a52 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 960.932671] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 960.932671] env[62753]: value = "task-1332484" [ 960.932671] env[62753]: _type = "Task" [ 960.932671] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 960.949207] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332484, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 960.994749] env[62753]: DEBUG oslo_vmware.rw_handles [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Completed reading data from the image iterator. {{(pid=62753) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 960.995093] env[62753]: DEBUG oslo_vmware.rw_handles [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Getting lease state for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e85ed3-0b7b-70a9-f231-32e43cafc227/disk-0.vmdk. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 960.996528] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e4c7aa0-c0a9-469e-9d4d-08673dbe2779 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.009937] env[62753]: DEBUG oslo_vmware.rw_handles [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e85ed3-0b7b-70a9-f231-32e43cafc227/disk-0.vmdk is in state: ready. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 961.010601] env[62753]: DEBUG oslo_vmware.rw_handles [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Releasing lease for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e85ed3-0b7b-70a9-f231-32e43cafc227/disk-0.vmdk. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 961.015048] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-ba7e4f84-654c-48d4-8bcc-5e7726d818c1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.017216] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332480, 'name': Destroy_Task, 'duration_secs': 0.497927} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.017565] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Destroyed the VM [ 961.017720] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Deleting Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 961.018473] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-bd37fd02-0844-4162-b68b-d14e8290729c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.027347] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 961.027347] env[62753]: value = "task-1332485" [ 961.027347] env[62753]: _type = "Task" [ 961.027347] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.038725] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332485, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.073412] env[62753]: DEBUG oslo_concurrency.lockutils [None req-de3f69a5-680c-418e-a226-142dc6a1992a tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "a401909c-5978-492b-a4fa-b17d38a5af44" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 7.791s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.171111] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Skipping network cache update for instance because it is being deleted. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9945}} [ 961.171767] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Skipping network cache update for instance because it is Building. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9941}} [ 961.171767] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.171767] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquired lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 961.171970] env[62753]: DEBUG nova.network.neutron [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Forcefully refreshing network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 961.171970] env[62753]: DEBUG nova.objects.instance [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lazy-loading 'info_cache' on Instance uuid 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 961.258080] env[62753]: DEBUG oslo_vmware.api [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Task: {'id': task-1332482, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.354705} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.259662] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.259891] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.261748] env[62753]: INFO nova.compute.claims [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 961.268355] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 961.268355] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 961.268355] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 961.268355] env[62753]: INFO nova.compute.manager [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Took 1.21 seconds to destroy the instance on the hypervisor. [ 961.268355] env[62753]: DEBUG oslo.service.loopingcall [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 961.268355] env[62753]: DEBUG nova.compute.manager [-] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 961.268355] env[62753]: DEBUG nova.network.neutron [-] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 961.271050] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 961.271050] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c60a5d-4376-adca-3113-efef5b3f4c7b" [ 961.271050] env[62753]: _type = "HttpNfcLease" [ 961.271050] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 961.302051] env[62753]: DEBUG oslo_vmware.rw_handles [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Closed VMDK write handle for https://esx7c1n3.openstack.eu-de-1.cloud.sap/nfc/52e85ed3-0b7b-70a9-f231-32e43cafc227/disk-0.vmdk. {{(pid=62753) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 961.302375] env[62753]: INFO nova.virt.vmwareapi.images [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Downloaded image file data 673e56af-bdea-458e-b52c-3abfd1149b15 [ 961.303425] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0b214b33-b859-45c4-a232-b99293d2cfc4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.326110] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b28a2e03-9ada-4f69-a3c8-b8103c688fab {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.336149] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "abd5ec6e-7b12-433b-9009-f9108ab4e318" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.336896] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "abd5ec6e-7b12-433b-9009-f9108ab4e318" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.336896] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "abd5ec6e-7b12-433b-9009-f9108ab4e318-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.337085] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "abd5ec6e-7b12-433b-9009-f9108ab4e318-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.339196] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "abd5ec6e-7b12-433b-9009-f9108ab4e318-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 961.343083] env[62753]: INFO nova.compute.manager [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Terminating instance [ 961.346012] env[62753]: DEBUG nova.compute.manager [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 961.346392] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 961.348081] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-93f2ecc6-7636-4071-b70e-859ef0493db7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.364309] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 961.364775] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-41d54df0-dcef-4030-8e97-dd2c657fed2d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.369545] env[62753]: INFO nova.virt.vmwareapi.images [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] The imported VM was unregistered [ 961.372107] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Caching image {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 961.372378] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Creating directory with path [datastore2] devstack-image-cache_base/673e56af-bdea-458e-b52c-3abfd1149b15 {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 961.374123] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9ab2af07-6d98-4fe3-884f-294a284fa999 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.376987] env[62753]: DEBUG oslo_vmware.api [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 961.376987] env[62753]: value = "task-1332487" [ 961.376987] env[62753]: _type = "Task" [ 961.376987] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.389155] env[62753]: DEBUG oslo_vmware.api [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332487, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.391295] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Created directory with path [datastore2] devstack-image-cache_base/673e56af-bdea-458e-b52c-3abfd1149b15 {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 961.391634] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_23afe077-5e3b-4381-907e-66033b929d0e/OSTACK_IMG_23afe077-5e3b-4381-907e-66033b929d0e.vmdk to [datastore2] devstack-image-cache_base/673e56af-bdea-458e-b52c-3abfd1149b15/673e56af-bdea-458e-b52c-3abfd1149b15.vmdk. {{(pid=62753) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 961.392354] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-49ca24fe-d4e9-4276-8f30-b8c42d070c25 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.402343] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 961.402343] env[62753]: value = "task-1332488" [ 961.402343] env[62753]: _type = "Task" [ 961.402343] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 961.414191] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332488, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.446026] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332484, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.540812] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332485, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.664308] env[62753]: DEBUG nova.compute.manager [req-fcee07a1-5644-427b-807a-a34cc4f27852 req-0cf0864f-f15f-41f8-89cb-8a80d0e4ead5 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Received event network-changed-7fb046f4-e470-4f6d-8fa3-73f3b0157020 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 961.664648] env[62753]: DEBUG nova.compute.manager [req-fcee07a1-5644-427b-807a-a34cc4f27852 req-0cf0864f-f15f-41f8-89cb-8a80d0e4ead5 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Refreshing instance network info cache due to event network-changed-7fb046f4-e470-4f6d-8fa3-73f3b0157020. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 961.664813] env[62753]: DEBUG oslo_concurrency.lockutils [req-fcee07a1-5644-427b-807a-a34cc4f27852 req-0cf0864f-f15f-41f8-89cb-8a80d0e4ead5 service nova] Acquiring lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 961.761265] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 961.761265] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c60a5d-4376-adca-3113-efef5b3f4c7b" [ 961.761265] env[62753]: _type = "HttpNfcLease" [ 961.761265] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 961.776883] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "bba5e3cb-c896-4ce7-ac7f-1cf2aba20465" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.776883] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "bba5e3cb-c896-4ce7-ac7f-1cf2aba20465" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.868474] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 961.868757] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 961.890646] env[62753]: DEBUG oslo_vmware.api [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332487, 'name': PowerOffVM_Task, 'duration_secs': 0.238058} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.890960] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 961.891168] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 961.891452] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-8c348023-af61-4dac-a456-c7f2471b8448 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 961.923560] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332488, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 961.952135] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332484, 'name': CreateSnapshot_Task, 'duration_secs': 0.818166} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 961.952720] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Created Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 961.954244] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31b5b12c-b097-49b1-8901-0c8c2c819969 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.042953] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332485, 'name': RemoveSnapshot_Task, 'duration_secs': 0.730945} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 962.044323] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Deleted Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 962.044323] env[62753]: DEBUG nova.compute.manager [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 962.044627] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-700bc603-11b5-4c8c-8afd-9edc672473ee {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.126148] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 962.126346] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 962.126589] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Deleting the datastore file [datastore2] abd5ec6e-7b12-433b-9009-f9108ab4e318 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 962.126873] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6da05a36-4839-4b64-86ba-42ced7b54e26 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.136745] env[62753]: DEBUG oslo_vmware.api [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 962.136745] env[62753]: value = "task-1332490" [ 962.136745] env[62753]: _type = "Task" [ 962.136745] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.148697] env[62753]: DEBUG oslo_vmware.api [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332490, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.262131] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 962.262131] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c60a5d-4376-adca-3113-efef5b3f4c7b" [ 962.262131] env[62753]: _type = "HttpNfcLease" [ 962.262131] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 962.283572] env[62753]: DEBUG nova.compute.manager [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 962.373341] env[62753]: DEBUG nova.compute.utils [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 962.418845] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332488, 'name': MoveVirtualDisk_Task} progress is 35%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.483024] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Creating linked-clone VM from snapshot {{(pid=62753) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 962.483024] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-d82207ef-c9e9-40f5-9522-dcd958b394c4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.496266] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 962.496266] env[62753]: value = "task-1332491" [ 962.496266] env[62753]: _type = "Task" [ 962.496266] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.507952] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332491, 'name': CloneVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.528368] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de689104-d8d5-43d2-9be1-899ce8649007 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.539391] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7317b76a-8109-41f2-8896-a1559d5e9550 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.582954] env[62753]: DEBUG nova.network.neutron [-] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 962.585486] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-428b26eb-d476-4284-9a4a-b5a388832576 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.590313] env[62753]: INFO nova.compute.manager [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Shelve offloading [ 962.594965] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 962.595724] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0e726a65-77fe-485a-9a67-a0621b6e3701 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.602495] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c66c120-1826-4c6e-8bc9-945318f8ae2d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.610836] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 962.610836] env[62753]: value = "task-1332492" [ 962.610836] env[62753]: _type = "Task" [ 962.610836] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 962.623813] env[62753]: DEBUG nova.compute.provider_tree [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 962.633599] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] VM already powered off {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 962.633825] env[62753]: DEBUG nova.compute.manager [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 962.635302] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da969944-72b4-4159-8c0f-53d9d97a8152 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 962.651025] env[62753]: DEBUG oslo_vmware.api [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332490, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 962.651025] env[62753]: DEBUG oslo_concurrency.lockutils [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "refresh_cache-628239e8-a447-4183-bed9-5550e53fda2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 962.651243] env[62753]: DEBUG oslo_concurrency.lockutils [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "refresh_cache-628239e8-a447-4183-bed9-5550e53fda2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 962.651656] env[62753]: DEBUG nova.network.neutron [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 962.765332] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 962.765332] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c60a5d-4376-adca-3113-efef5b3f4c7b" [ 962.765332] env[62753]: _type = "HttpNfcLease" [ 962.765332] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 962.806821] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 962.880974] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.011s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 962.917785] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332488, 'name': MoveVirtualDisk_Task} progress is 57%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.006731] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332491, 'name': CloneVM_Task} progress is 93%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.019487] env[62753]: DEBUG nova.network.neutron [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Updating instance_info_cache with network_info: [{"id": "7fb046f4-e470-4f6d-8fa3-73f3b0157020", "address": "fa:16:3e:6f:26:2c", "network": {"id": "fc646f7f-70ba-4671-8476-3b0ece1e9041", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-853174775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d0eb0b3ba504f59a85cb021da2b47b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fb046f4-e4", "ovs_interfaceid": "7fb046f4-e470-4f6d-8fa3-73f3b0157020", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.048110] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "a401909c-5978-492b-a4fa-b17d38a5af44" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.048357] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "a401909c-5978-492b-a4fa-b17d38a5af44" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.048553] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "a401909c-5978-492b-a4fa-b17d38a5af44-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.049652] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "a401909c-5978-492b-a4fa-b17d38a5af44-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.049652] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "a401909c-5978-492b-a4fa-b17d38a5af44-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.051312] env[62753]: INFO nova.compute.manager [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Terminating instance [ 963.053289] env[62753]: DEBUG nova.compute.manager [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 963.053540] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 963.054597] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a81eefa-b046-42c8-bd88-06f384320754 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.065704] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 963.066015] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6d64fdaa-4720-4d5a-92b0-98a1be37cc08 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.076906] env[62753]: DEBUG oslo_vmware.api [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 963.076906] env[62753]: value = "task-1332493" [ 963.076906] env[62753]: _type = "Task" [ 963.076906] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 963.089066] env[62753]: DEBUG oslo_vmware.api [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332493, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.093846] env[62753]: INFO nova.compute.manager [-] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Took 1.83 seconds to deallocate network for instance. [ 963.127388] env[62753]: DEBUG nova.scheduler.client.report [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 963.157063] env[62753]: DEBUG oslo_vmware.api [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332490, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.260094] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 963.260094] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c60a5d-4376-adca-3113-efef5b3f4c7b" [ 963.260094] env[62753]: _type = "HttpNfcLease" [ 963.260094] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 963.409468] env[62753]: DEBUG nova.network.neutron [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Updating instance_info_cache with network_info: [{"id": "800db109-09e3-45a4-8753-20d9911c0677", "address": "fa:16:3e:0e:c9:9f", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap800db109-09", "ovs_interfaceid": "800db109-09e3-45a4-8753-20d9911c0677", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 963.419946] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332488, 'name': MoveVirtualDisk_Task} progress is 77%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.509505] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332491, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.522875] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Releasing lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.523113] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Updated the network info_cache for instance {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 963.523408] env[62753]: DEBUG oslo_concurrency.lockutils [req-fcee07a1-5644-427b-807a-a34cc4f27852 req-0cf0864f-f15f-41f8-89cb-8a80d0e4ead5 service nova] Acquired lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 963.523673] env[62753]: DEBUG nova.network.neutron [req-fcee07a1-5644-427b-807a-a34cc4f27852 req-0cf0864f-f15f-41f8-89cb-8a80d0e4ead5 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Refreshing network info cache for port 7fb046f4-e470-4f6d-8fa3-73f3b0157020 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 963.591352] env[62753]: DEBUG oslo_vmware.api [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332493, 'name': PowerOffVM_Task, 'duration_secs': 0.237956} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 963.591732] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 963.591985] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 963.592298] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-44ffd013-492c-4125-9024-9fc59b713b72 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.601105] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.633481] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.373s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 963.634239] env[62753]: DEBUG nova.compute.manager [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 963.637573] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.831s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.639354] env[62753]: INFO nova.compute.claims [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 963.655659] env[62753]: DEBUG oslo_vmware.api [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332490, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.760698] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 963.760698] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c60a5d-4376-adca-3113-efef5b3f4c7b" [ 963.760698] env[62753]: _type = "HttpNfcLease" [ 963.760698] env[62753]: } is ready. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 963.761081] env[62753]: DEBUG oslo_vmware.rw_handles [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 963.761081] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c60a5d-4376-adca-3113-efef5b3f4c7b" [ 963.761081] env[62753]: _type = "HttpNfcLease" [ 963.761081] env[62753]: }. {{(pid=62753) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 963.761961] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1882d8c-071e-4d85-a0a2-1ef47ad776be {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.771062] env[62753]: DEBUG oslo_vmware.rw_handles [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fcf6ca-e803-b989-90f4-c5a036ef7820/disk-0.vmdk from lease info. {{(pid=62753) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 963.771440] env[62753]: DEBUG oslo_vmware.rw_handles [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Creating HTTP connection to write to file with size = 31667712 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fcf6ca-e803-b989-90f4-c5a036ef7820/disk-0.vmdk. {{(pid=62753) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 963.839209] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9bd737f0-0840-41ff-9cf7-febb12de4596 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 963.915456] env[62753]: DEBUG oslo_concurrency.lockutils [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "refresh_cache-628239e8-a447-4183-bed9-5550e53fda2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 963.917462] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332488, 'name': MoveVirtualDisk_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 963.919700] env[62753]: DEBUG nova.compute.manager [req-478fc0a9-f27c-4bc7-a719-385b385b1318 req-45c4960c-336e-4845-b975-04387e5f6f26 service nova] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Received event network-vif-deleted-ef993aa6-10f5-4de9-b652-426dbaac5112 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 963.951086] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 963.951345] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 963.951628] env[62753]: INFO nova.compute.manager [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Attaching volume 36f91b11-b0aa-4304-aa25-a7ff2fa37218 to /dev/sdc [ 963.989278] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-172a1e80-6708-401c-8357-632801d4208f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.001269] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72d29bc9-c296-4fb5-b4df-3907c31855bd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.010364] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332491, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.016305] env[62753]: DEBUG nova.virt.block_device [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Updating existing volume attachment record: b9eae3f1-21b6-459a-9712-dda95937107e {{(pid=62753) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 964.086087] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 964.086390] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 964.086646] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Deleting the datastore file [datastore1] a401909c-5978-492b-a4fa-b17d38a5af44 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.088296] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-54ecd4bb-5dc9-4afe-ab66-060569f4be63 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.096951] env[62753]: DEBUG oslo_vmware.api [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 964.096951] env[62753]: value = "task-1332495" [ 964.096951] env[62753]: _type = "Task" [ 964.096951] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.109582] env[62753]: DEBUG oslo_vmware.api [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332495, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.148589] env[62753]: DEBUG nova.compute.utils [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 964.154966] env[62753]: DEBUG nova.compute.manager [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 964.155250] env[62753]: DEBUG nova.network.neutron [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 964.179121] env[62753]: DEBUG oslo_vmware.api [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332490, 'name': DeleteDatastoreFile_Task, 'duration_secs': 1.774186} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.179323] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 964.179665] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 964.180012] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 964.180363] env[62753]: INFO nova.compute.manager [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Took 2.83 seconds to destroy the instance on the hypervisor. [ 964.180810] env[62753]: DEBUG oslo.service.loopingcall [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 964.184282] env[62753]: DEBUG nova.compute.manager [-] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 964.184471] env[62753]: DEBUG nova.network.neutron [-] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 964.253595] env[62753]: DEBUG nova.policy [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9931b40cfd7846038805c6a4caedaac0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b8e81660b30408c998e412f5fa81469', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 964.406224] env[62753]: DEBUG nova.network.neutron [req-fcee07a1-5644-427b-807a-a34cc4f27852 req-0cf0864f-f15f-41f8-89cb-8a80d0e4ead5 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Updated VIF entry in instance network info cache for port 7fb046f4-e470-4f6d-8fa3-73f3b0157020. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 964.406720] env[62753]: DEBUG nova.network.neutron [req-fcee07a1-5644-427b-807a-a34cc4f27852 req-0cf0864f-f15f-41f8-89cb-8a80d0e4ead5 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Updating instance_info_cache with network_info: [{"id": "7fb046f4-e470-4f6d-8fa3-73f3b0157020", "address": "fa:16:3e:6f:26:2c", "network": {"id": "fc646f7f-70ba-4671-8476-3b0ece1e9041", "bridge": "br-int", "label": "tempest-AttachVolumeShelveTestJSON-853174775-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.163", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "4d0eb0b3ba504f59a85cb021da2b47b2", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "255460d5-71d4-4bfd-87f1-acc10085db7f", "external-id": "nsx-vlan-transportzone-152", "segmentation_id": 152, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7fb046f4-e4", "ovs_interfaceid": "7fb046f4-e470-4f6d-8fa3-73f3b0157020", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 964.417618] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 964.418649] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de92f90f-6a3d-4794-a40f-7cb35bb30fb8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.426340] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332488, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.595442} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.426645] env[62753]: INFO nova.virt.vmwareapi.ds_util [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_23afe077-5e3b-4381-907e-66033b929d0e/OSTACK_IMG_23afe077-5e3b-4381-907e-66033b929d0e.vmdk to [datastore2] devstack-image-cache_base/673e56af-bdea-458e-b52c-3abfd1149b15/673e56af-bdea-458e-b52c-3abfd1149b15.vmdk. [ 964.427392] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Cleaning up location [datastore2] OSTACK_IMG_23afe077-5e3b-4381-907e-66033b929d0e {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 964.427733] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_23afe077-5e3b-4381-907e-66033b929d0e {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.428347] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b714111b-b0e3-4e5c-9021-c47c6ddf900d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.435357] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 964.436531] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-33a451d3-9a15-464f-8441-99946898185b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.440870] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 964.440870] env[62753]: value = "task-1332497" [ 964.440870] env[62753]: _type = "Task" [ 964.440870] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.449752] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332497, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.515160] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332491, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.518509] env[62753]: DEBUG oslo_vmware.rw_handles [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527203da-b726-a702-98cc-f84eb8d56997/disk-0.vmdk. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 964.519568] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 964.519568] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 964.519568] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Deleting the datastore file [datastore2] 628239e8-a447-4183-bed9-5550e53fda2b {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 964.520033] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69e1b26c-48d9-41eb-ac28-ea09b541b802 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.524293] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-3f6715a9-3736-48be-ad9e-32e1f5cdb357 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.531381] env[62753]: DEBUG oslo_vmware.rw_handles [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527203da-b726-a702-98cc-f84eb8d56997/disk-0.vmdk is in state: ready. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 964.531606] env[62753]: ERROR oslo_vmware.rw_handles [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527203da-b726-a702-98cc-f84eb8d56997/disk-0.vmdk due to incomplete transfer. [ 964.533853] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-f5779fcd-6f63-49c5-a968-ed912069e846 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.535771] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 964.535771] env[62753]: value = "task-1332499" [ 964.535771] env[62753]: _type = "Task" [ 964.535771] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.547508] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332499, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.551315] env[62753]: DEBUG oslo_vmware.rw_handles [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/527203da-b726-a702-98cc-f84eb8d56997/disk-0.vmdk. {{(pid=62753) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 964.551557] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Uploaded image 03c3734b-05a6-4c37-9f4e-406198dd6413 to the Glance image server {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 964.554116] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Destroying the VM {{(pid=62753) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 964.554418] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-9e4722a5-544f-46d6-a3a6-139666c12d39 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.566750] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 964.566750] env[62753]: value = "task-1332500" [ 964.566750] env[62753]: _type = "Task" [ 964.566750] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 964.580530] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332500, 'name': Destroy_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 964.611805] env[62753]: DEBUG oslo_vmware.api [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332495, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.281} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.614229] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 964.614462] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 964.614654] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 964.614840] env[62753]: INFO nova.compute.manager [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Took 1.56 seconds to destroy the instance on the hypervisor. [ 964.615137] env[62753]: DEBUG oslo.service.loopingcall [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 964.615734] env[62753]: DEBUG nova.compute.manager [-] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 964.615832] env[62753]: DEBUG nova.network.neutron [-] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 964.618706] env[62753]: DEBUG nova.network.neutron [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Successfully created port: 0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 964.655800] env[62753]: DEBUG nova.compute.manager [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 964.916971] env[62753]: DEBUG oslo_concurrency.lockutils [req-fcee07a1-5644-427b-807a-a34cc4f27852 req-0cf0864f-f15f-41f8-89cb-8a80d0e4ead5 service nova] Releasing lock "refresh_cache-4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.949460] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e50afca-324e-47ca-8793-bda4bac270cb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.964524] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e010f7fa-00a8-489d-931b-86ecb7d34a60 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 964.968278] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332497, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.060262} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 964.971346] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 964.971695] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Releasing lock "[datastore2] devstack-image-cache_base/673e56af-bdea-458e-b52c-3abfd1149b15/673e56af-bdea-458e-b52c-3abfd1149b15.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 964.971900] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/673e56af-bdea-458e-b52c-3abfd1149b15/673e56af-bdea-458e-b52c-3abfd1149b15.vmdk to [datastore2] ff4b295e-2eb2-42dd-b353-08df755f04ea/ff4b295e-2eb2-42dd-b353-08df755f04ea.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 964.972680] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6393f148-ca1f-4bdb-9349-cf5a88d94893 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.011245] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d1b2d3-d03f-4c38-97e7-06e6347b0d24 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.013936] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 965.013936] env[62753]: value = "task-1332501" [ 965.013936] env[62753]: _type = "Task" [ 965.013936] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.026718] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332491, 'name': CloneVM_Task, 'duration_secs': 2.431914} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.029393] env[62753]: INFO nova.virt.vmwareapi.vmops [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Created linked-clone VM from snapshot [ 965.030685] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f44acca-ea7b-47b5-a763-122ae6b24eaf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.038590] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-235acfc0-e353-486e-9db5-1afdb9714df4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.041570] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332501, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.055681] env[62753]: DEBUG nova.compute.provider_tree [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 965.063543] env[62753]: DEBUG oslo_vmware.rw_handles [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Completed reading data from the image iterator. {{(pid=62753) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 965.063543] env[62753]: DEBUG oslo_vmware.rw_handles [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fcf6ca-e803-b989-90f4-c5a036ef7820/disk-0.vmdk. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 965.063999] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Uploading image 3b2f7019-a7ab-4e0e-9ea9-07c8a7334614 {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 965.066853] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d4020b9-c489-4fd9-83c2-ca8d449b8b63 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.074351] env[62753]: DEBUG nova.scheduler.client.report [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 965.078031] env[62753]: DEBUG oslo_vmware.api [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332499, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.279062} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.078305] env[62753]: DEBUG nova.network.neutron [-] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.085153] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 965.085153] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 965.085153] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 965.088811] env[62753]: DEBUG oslo_vmware.rw_handles [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fcf6ca-e803-b989-90f4-c5a036ef7820/disk-0.vmdk is in state: ready. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 965.088978] env[62753]: DEBUG oslo_vmware.rw_handles [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fcf6ca-e803-b989-90f4-c5a036ef7820/disk-0.vmdk. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 965.089640] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-45016c37-6046-42ba-9730-453c0ecab857 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.094913] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332500, 'name': Destroy_Task, 'duration_secs': 0.454287} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 965.095640] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Destroyed the VM [ 965.095881] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Deleting Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 965.096753] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-9fafe5b6-0fc7-404a-9eb4-e65e9863e912 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.106571] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 965.106571] env[62753]: value = "task-1332502" [ 965.106571] env[62753]: _type = "Task" [ 965.106571] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.109492] env[62753]: DEBUG oslo_vmware.rw_handles [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 965.109492] env[62753]: value = "vm-284642" [ 965.109492] env[62753]: _type = "VirtualMachine" [ 965.109492] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 965.110144] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-1e9e9fbf-6e1d-478f-b239-d955696a5e80 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.117474] env[62753]: INFO nova.scheduler.client.report [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Deleted allocations for instance 628239e8-a447-4183-bed9-5550e53fda2b [ 965.127507] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332502, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.130452] env[62753]: DEBUG oslo_vmware.rw_handles [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lease: (returnval){ [ 965.130452] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f41fbb-3ec1-aed7-5ecd-2636e97ce875" [ 965.130452] env[62753]: _type = "HttpNfcLease" [ 965.130452] env[62753]: } obtained for exporting VM: (result){ [ 965.130452] env[62753]: value = "vm-284642" [ 965.130452] env[62753]: _type = "VirtualMachine" [ 965.130452] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 965.130717] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the lease: (returnval){ [ 965.130717] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f41fbb-3ec1-aed7-5ecd-2636e97ce875" [ 965.130717] env[62753]: _type = "HttpNfcLease" [ 965.130717] env[62753]: } to be ready. {{(pid=62753) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 965.137681] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 965.137681] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f41fbb-3ec1-aed7-5ecd-2636e97ce875" [ 965.137681] env[62753]: _type = "HttpNfcLease" [ 965.137681] env[62753]: } is ready. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 965.137966] env[62753]: DEBUG oslo_vmware.rw_handles [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 965.137966] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f41fbb-3ec1-aed7-5ecd-2636e97ce875" [ 965.137966] env[62753]: _type = "HttpNfcLease" [ 965.137966] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 965.138757] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d3134a4-608d-4176-b37e-dd7b035944d1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.149905] env[62753]: DEBUG oslo_vmware.rw_handles [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52669950-1499-f8bd-49c0-ae70a3b18c65/disk-0.vmdk from lease info. {{(pid=62753) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 965.150138] env[62753]: DEBUG oslo_vmware.rw_handles [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52669950-1499-f8bd-49c0-ae70a3b18c65/disk-0.vmdk for reading. {{(pid=62753) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 965.313816] env[62753]: DEBUG oslo_vmware.rw_handles [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/52fcf6ca-e803-b989-90f4-c5a036ef7820/disk-0.vmdk. {{(pid=62753) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 965.314178] env[62753]: INFO nova.virt.vmwareapi.images [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Downloaded image file data 061f0060-350a-4bf7-a725-58acba59de2b [ 965.315274] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f7dff1e-52ae-4a62-aa34-8a8d46beb67a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.337167] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-86d8a8de-7c5c-47c3-9d9a-8216fecba837 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.381942] env[62753]: INFO nova.virt.vmwareapi.images [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] The imported VM was unregistered [ 965.385346] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Caching image {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 965.385716] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Creating directory with path [datastore1] devstack-image-cache_base/061f0060-350a-4bf7-a725-58acba59de2b {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 965.386785] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-884954c1-aab1-4106-b02c-e2e74a355421 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.401446] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Created directory with path [datastore1] devstack-image-cache_base/061f0060-350a-4bf7-a725-58acba59de2b {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 965.401716] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Moving virtual disk from [datastore1] OSTACK_IMG_715833fa-e566-42a9-8715-2235ec81fd4d/OSTACK_IMG_715833fa-e566-42a9-8715-2235ec81fd4d.vmdk to [datastore1] devstack-image-cache_base/061f0060-350a-4bf7-a725-58acba59de2b/061f0060-350a-4bf7-a725-58acba59de2b.vmdk. {{(pid=62753) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 965.402228] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-e87da92a-9efe-46b6-b17d-4827c268b193 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.411257] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 965.411257] env[62753]: value = "task-1332505" [ 965.411257] env[62753]: _type = "Task" [ 965.411257] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 965.422867] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332505, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.529319] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332501, 'name': CopyVirtualDisk_Task} progress is 24%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.552602] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.552810] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.559849] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-adf14e1f-761f-4cc7-878f-32b2d526651a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.585627] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 1.948s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 965.586365] env[62753]: DEBUG nova.compute.manager [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 965.591029] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.989s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 965.591029] env[62753]: DEBUG nova.objects.instance [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Lazy-loading 'resources' on Instance uuid d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 965.592230] env[62753]: INFO nova.compute.manager [-] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Took 1.41 seconds to deallocate network for instance. [ 965.623282] env[62753]: DEBUG oslo_concurrency.lockutils [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 965.623980] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332502, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 965.731764] env[62753]: DEBUG nova.compute.manager [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 965.770195] env[62753]: DEBUG nova.virt.hardware [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 965.770581] env[62753]: DEBUG nova.virt.hardware [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 965.770788] env[62753]: DEBUG nova.virt.hardware [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 965.771084] env[62753]: DEBUG nova.virt.hardware [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 965.771653] env[62753]: DEBUG nova.virt.hardware [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 965.772153] env[62753]: DEBUG nova.virt.hardware [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 965.772455] env[62753]: DEBUG nova.virt.hardware [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 965.772724] env[62753]: DEBUG nova.virt.hardware [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 965.773214] env[62753]: DEBUG nova.virt.hardware [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 965.773693] env[62753]: DEBUG nova.virt.hardware [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 965.773941] env[62753]: DEBUG nova.virt.hardware [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 965.775145] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74f08c3d-7268-4093-b083-969e2788234c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.793497] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-50d08eae-ea86-4a3b-9613-95b8a14cccdb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 965.876510] env[62753]: DEBUG nova.network.neutron [-] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 965.927349] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332505, 'name': MoveVirtualDisk_Task} progress is 15%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.019273] env[62753]: DEBUG nova.compute.manager [req-5544b7d2-f5a6-4805-b3f4-8d6656396c0b req-e2b2957e-48ac-4040-9b95-d195e1dab196 service nova] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Received event network-vif-unplugged-800db109-09e3-45a4-8753-20d9911c0677 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.019273] env[62753]: DEBUG oslo_concurrency.lockutils [req-5544b7d2-f5a6-4805-b3f4-8d6656396c0b req-e2b2957e-48ac-4040-9b95-d195e1dab196 service nova] Acquiring lock "628239e8-a447-4183-bed9-5550e53fda2b-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.019433] env[62753]: DEBUG oslo_concurrency.lockutils [req-5544b7d2-f5a6-4805-b3f4-8d6656396c0b req-e2b2957e-48ac-4040-9b95-d195e1dab196 service nova] Lock "628239e8-a447-4183-bed9-5550e53fda2b-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.019598] env[62753]: DEBUG oslo_concurrency.lockutils [req-5544b7d2-f5a6-4805-b3f4-8d6656396c0b req-e2b2957e-48ac-4040-9b95-d195e1dab196 service nova] Lock "628239e8-a447-4183-bed9-5550e53fda2b-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.019777] env[62753]: DEBUG nova.compute.manager [req-5544b7d2-f5a6-4805-b3f4-8d6656396c0b req-e2b2957e-48ac-4040-9b95-d195e1dab196 service nova] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] No waiting events found dispatching network-vif-unplugged-800db109-09e3-45a4-8753-20d9911c0677 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 966.019947] env[62753]: DEBUG nova.compute.manager [req-5544b7d2-f5a6-4805-b3f4-8d6656396c0b req-e2b2957e-48ac-4040-9b95-d195e1dab196 service nova] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Received event network-vif-unplugged-800db109-09e3-45a4-8753-20d9911c0677 for instance with task_state deleting. {{(pid=62753) _process_instance_event /opt/stack/nova/nova/compute/manager.py:10909}} [ 966.020203] env[62753]: DEBUG nova.compute.manager [req-5544b7d2-f5a6-4805-b3f4-8d6656396c0b req-e2b2957e-48ac-4040-9b95-d195e1dab196 service nova] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Received event network-changed-800db109-09e3-45a4-8753-20d9911c0677 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.020304] env[62753]: DEBUG nova.compute.manager [req-5544b7d2-f5a6-4805-b3f4-8d6656396c0b req-e2b2957e-48ac-4040-9b95-d195e1dab196 service nova] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Refreshing instance network info cache due to event network-changed-800db109-09e3-45a4-8753-20d9911c0677. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 966.020498] env[62753]: DEBUG oslo_concurrency.lockutils [req-5544b7d2-f5a6-4805-b3f4-8d6656396c0b req-e2b2957e-48ac-4040-9b95-d195e1dab196 service nova] Acquiring lock "refresh_cache-628239e8-a447-4183-bed9-5550e53fda2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.020637] env[62753]: DEBUG oslo_concurrency.lockutils [req-5544b7d2-f5a6-4805-b3f4-8d6656396c0b req-e2b2957e-48ac-4040-9b95-d195e1dab196 service nova] Acquired lock "refresh_cache-628239e8-a447-4183-bed9-5550e53fda2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.020820] env[62753]: DEBUG nova.network.neutron [req-5544b7d2-f5a6-4805-b3f4-8d6656396c0b req-e2b2957e-48ac-4040-9b95-d195e1dab196 service nova] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Refreshing network info cache for port 800db109-09e3-45a4-8753-20d9911c0677 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 966.039065] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332501, 'name': CopyVirtualDisk_Task} progress is 46%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.055637] env[62753]: DEBUG nova.compute.manager [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 966.093863] env[62753]: DEBUG nova.compute.utils [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 966.097173] env[62753]: DEBUG nova.compute.manager [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 966.097173] env[62753]: DEBUG nova.network.neutron [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 966.105708] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.120916] env[62753]: DEBUG nova.compute.manager [req-44ea541b-6b15-4b87-ad71-0282060c7ab8 req-44bde087-1fa5-4d35-a3c9-5fe958c7369c service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Received event network-vif-plugged-0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 966.121341] env[62753]: DEBUG oslo_concurrency.lockutils [req-44ea541b-6b15-4b87-ad71-0282060c7ab8 req-44bde087-1fa5-4d35-a3c9-5fe958c7369c service nova] Acquiring lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.121610] env[62753]: DEBUG oslo_concurrency.lockutils [req-44ea541b-6b15-4b87-ad71-0282060c7ab8 req-44bde087-1fa5-4d35-a3c9-5fe958c7369c service nova] Lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 966.121810] env[62753]: DEBUG oslo_concurrency.lockutils [req-44ea541b-6b15-4b87-ad71-0282060c7ab8 req-44bde087-1fa5-4d35-a3c9-5fe958c7369c service nova] Lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 966.122120] env[62753]: DEBUG nova.compute.manager [req-44ea541b-6b15-4b87-ad71-0282060c7ab8 req-44bde087-1fa5-4d35-a3c9-5fe958c7369c service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] No waiting events found dispatching network-vif-plugged-0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 966.122347] env[62753]: WARNING nova.compute.manager [req-44ea541b-6b15-4b87-ad71-0282060c7ab8 req-44bde087-1fa5-4d35-a3c9-5fe958c7369c service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Received unexpected event network-vif-plugged-0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9 for instance with vm_state building and task_state spawning. [ 966.127448] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332502, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.153876] env[62753]: DEBUG nova.policy [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6996040b35454e028b94735be079b925', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37368e15fa5a4a0a97ac1c22fcf0bb80', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 966.241393] env[62753]: DEBUG nova.network.neutron [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Successfully updated port: 0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 966.379967] env[62753]: INFO nova.compute.manager [-] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Took 1.76 seconds to deallocate network for instance. [ 966.399419] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb1caa3d-2623-4864-9a09-dc5721df5eb7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.413353] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb9c48c9-5562-468c-aa04-4244e8a21f52 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.429058] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332505, 'name': MoveVirtualDisk_Task} progress is 38%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.455939] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f37d027a-f958-46b1-80be-03bde5e3df33 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.467991] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82678539-b4ee-4c0b-bd06-b27e4d6c3b3d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 966.491734] env[62753]: DEBUG nova.compute.provider_tree [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 966.497036] env[62753]: DEBUG oslo_concurrency.lockutils [None req-de351a37-b138-4888-adf5-f77f8aa16990 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "628239e8-a447-4183-bed9-5550e53fda2b" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.540579] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332501, 'name': CopyVirtualDisk_Task} progress is 71%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.553212] env[62753]: DEBUG nova.network.neutron [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Successfully created port: 4832b8bd-7a18-4f8e-a05b-7421417d64ff {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 966.581426] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.605069] env[62753]: DEBUG nova.compute.manager [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 966.627498] env[62753]: DEBUG oslo_vmware.api [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332502, 'name': RemoveSnapshot_Task, 'duration_secs': 1.421679} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 966.627793] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Deleted Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 966.628099] env[62753]: INFO nova.compute.manager [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Took 16.77 seconds to snapshot the instance on the hypervisor. [ 966.747115] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 966.747369] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 966.747424] env[62753]: DEBUG nova.network.neutron [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 966.763225] env[62753]: DEBUG nova.network.neutron [req-5544b7d2-f5a6-4805-b3f4-8d6656396c0b req-e2b2957e-48ac-4040-9b95-d195e1dab196 service nova] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Updated VIF entry in instance network info cache for port 800db109-09e3-45a4-8753-20d9911c0677. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 966.763685] env[62753]: DEBUG nova.network.neutron [req-5544b7d2-f5a6-4805-b3f4-8d6656396c0b req-e2b2957e-48ac-4040-9b95-d195e1dab196 service nova] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Updating instance_info_cache with network_info: [{"id": "800db109-09e3-45a4-8753-20d9911c0677", "address": "fa:16:3e:0e:c9:9f", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": null, "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap800db109-09", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 966.893326] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 966.939291] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332505, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 966.995204] env[62753]: DEBUG nova.scheduler.client.report [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 967.038895] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332501, 'name': CopyVirtualDisk_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.268025] env[62753]: DEBUG oslo_concurrency.lockutils [req-5544b7d2-f5a6-4805-b3f4-8d6656396c0b req-e2b2957e-48ac-4040-9b95-d195e1dab196 service nova] Releasing lock "refresh_cache-628239e8-a447-4183-bed9-5550e53fda2b" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 967.268025] env[62753]: DEBUG nova.compute.manager [req-5544b7d2-f5a6-4805-b3f4-8d6656396c0b req-e2b2957e-48ac-4040-9b95-d195e1dab196 service nova] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Received event network-vif-deleted-a98dff4a-a9af-4961-8146-fbd28735e7af {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.268025] env[62753]: DEBUG nova.compute.manager [req-5544b7d2-f5a6-4805-b3f4-8d6656396c0b req-e2b2957e-48ac-4040-9b95-d195e1dab196 service nova] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Received event network-vif-deleted-7df5849e-e6d9-4ee9-92eb-9b64c9b9cc36 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 967.306150] env[62753]: DEBUG nova.network.neutron [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 967.431810] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332505, 'name': MoveVirtualDisk_Task} progress is 85%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.498409] env[62753]: DEBUG nova.compute.manager [None req-fd27cc6b-e9e2-446c-9f08-857552426d7a tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Found 2 images (rotation: 2) {{(pid=62753) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 967.501926] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.911s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 967.505251] env[62753]: DEBUG oslo_concurrency.lockutils [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.882s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 967.505251] env[62753]: DEBUG nova.objects.instance [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lazy-loading 'resources' on Instance uuid 628239e8-a447-4183-bed9-5550e53fda2b {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 967.509219] env[62753]: DEBUG nova.network.neutron [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Updating instance_info_cache with network_info: [{"id": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "address": "fa:16:3e:6b:34:1e", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cd05bfa-82", "ovs_interfaceid": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 967.536260] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332501, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.184237} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.536329] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/673e56af-bdea-458e-b52c-3abfd1149b15/673e56af-bdea-458e-b52c-3abfd1149b15.vmdk to [datastore2] ff4b295e-2eb2-42dd-b353-08df755f04ea/ff4b295e-2eb2-42dd-b353-08df755f04ea.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 967.537762] env[62753]: INFO nova.scheduler.client.report [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Deleted allocations for instance d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075 [ 967.539569] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef703389-71df-4d76-a806-87273d0f6eec {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.572875] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Reconfiguring VM instance instance-00000053 to attach disk [datastore2] ff4b295e-2eb2-42dd-b353-08df755f04ea/ff4b295e-2eb2-42dd-b353-08df755f04ea.vmdk or device None with type streamOptimized {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 967.573379] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f5a50383-1170-4af6-a013-9dcf5bddee9d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.597453] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 967.597453] env[62753]: value = "task-1332507" [ 967.597453] env[62753]: _type = "Task" [ 967.597453] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.612725] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332507, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.619718] env[62753]: DEBUG nova.compute.manager [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 967.655825] env[62753]: DEBUG nova.virt.hardware [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 967.655825] env[62753]: DEBUG nova.virt.hardware [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 967.655825] env[62753]: DEBUG nova.virt.hardware [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 967.655825] env[62753]: DEBUG nova.virt.hardware [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 967.655825] env[62753]: DEBUG nova.virt.hardware [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 967.656932] env[62753]: DEBUG nova.virt.hardware [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 967.656932] env[62753]: DEBUG nova.virt.hardware [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 967.656932] env[62753]: DEBUG nova.virt.hardware [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 967.657246] env[62753]: DEBUG nova.virt.hardware [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 967.657440] env[62753]: DEBUG nova.virt.hardware [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 967.658043] env[62753]: DEBUG nova.virt.hardware [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 967.658658] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-26e3e279-7d9d-4800-b2cc-b99d000f16c5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.669650] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6565f6ab-fc76-4061-9b77-a06ba5c1bbfe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.930404] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332505, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.459858} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 967.931008] env[62753]: INFO nova.virt.vmwareapi.ds_util [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Moved virtual disk from [datastore1] OSTACK_IMG_715833fa-e566-42a9-8715-2235ec81fd4d/OSTACK_IMG_715833fa-e566-42a9-8715-2235ec81fd4d.vmdk to [datastore1] devstack-image-cache_base/061f0060-350a-4bf7-a725-58acba59de2b/061f0060-350a-4bf7-a725-58acba59de2b.vmdk. [ 967.931262] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Cleaning up location [datastore1] OSTACK_IMG_715833fa-e566-42a9-8715-2235ec81fd4d {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 967.931499] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Deleting the datastore file [datastore1] OSTACK_IMG_715833fa-e566-42a9-8715-2235ec81fd4d {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 967.931887] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4de380e3-1613-4962-b121-d612578a9614 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 967.941214] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 967.941214] env[62753]: value = "task-1332508" [ 967.941214] env[62753]: _type = "Task" [ 967.941214] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 967.950530] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332508, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 967.991574] env[62753]: DEBUG nova.compute.manager [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 967.992743] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c02ae077-d67c-4945-843c-4dab1dacf222 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.008416] env[62753]: DEBUG nova.objects.instance [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lazy-loading 'numa_topology' on Instance uuid 628239e8-a447-4183-bed9-5550e53fda2b {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.016626] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.016991] env[62753]: DEBUG nova.compute.manager [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Instance network_info: |[{"id": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "address": "fa:16:3e:6b:34:1e", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cd05bfa-82", "ovs_interfaceid": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 968.017518] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6b:34:1e', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'd48f0ef6-34e5-44d4-8baf-4470ed96ce73', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 968.026539] env[62753]: DEBUG oslo.service.loopingcall [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 968.027159] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 968.027518] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c42c29c9-f278-4c13-a568-05cd68ba198d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.049356] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dc63f225-9d4e-41e8-8e35-43e4db23b2d9 tempest-ServersV294TestFqdnHostnames-1249078754 tempest-ServersV294TestFqdnHostnames-1249078754-project-member] Lock "d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.999s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.054303] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 968.054303] env[62753]: value = "task-1332509" [ 968.054303] env[62753]: _type = "Task" [ 968.054303] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.066068] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332509, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.110921] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332507, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.152821] env[62753]: DEBUG nova.compute.manager [req-948a5f5d-7d93-47a7-82bd-46793ccb1607 req-d0a9366e-06f9-4bea-9c19-a037f50fdb73 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Received event network-changed-0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 968.153534] env[62753]: DEBUG nova.compute.manager [req-948a5f5d-7d93-47a7-82bd-46793ccb1607 req-d0a9366e-06f9-4bea-9c19-a037f50fdb73 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Refreshing instance network info cache due to event network-changed-0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 968.153779] env[62753]: DEBUG oslo_concurrency.lockutils [req-948a5f5d-7d93-47a7-82bd-46793ccb1607 req-d0a9366e-06f9-4bea-9c19-a037f50fdb73 service nova] Acquiring lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.153970] env[62753]: DEBUG oslo_concurrency.lockutils [req-948a5f5d-7d93-47a7-82bd-46793ccb1607 req-d0a9366e-06f9-4bea-9c19-a037f50fdb73 service nova] Acquired lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.154200] env[62753]: DEBUG nova.network.neutron [req-948a5f5d-7d93-47a7-82bd-46793ccb1607 req-d0a9366e-06f9-4bea-9c19-a037f50fdb73 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Refreshing network info cache for port 0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 968.361192] env[62753]: DEBUG nova.compute.manager [req-bb1f7144-fe43-4cde-bc29-b6d2c0ab067f req-bcdbad83-2934-49e4-bc56-5d09f22e22ee service nova] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Received event network-vif-plugged-4832b8bd-7a18-4f8e-a05b-7421417d64ff {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 968.361664] env[62753]: DEBUG oslo_concurrency.lockutils [req-bb1f7144-fe43-4cde-bc29-b6d2c0ab067f req-bcdbad83-2934-49e4-bc56-5d09f22e22ee service nova] Acquiring lock "bba5e3cb-c896-4ce7-ac7f-1cf2aba20465-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 968.362122] env[62753]: DEBUG oslo_concurrency.lockutils [req-bb1f7144-fe43-4cde-bc29-b6d2c0ab067f req-bcdbad83-2934-49e4-bc56-5d09f22e22ee service nova] Lock "bba5e3cb-c896-4ce7-ac7f-1cf2aba20465-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 968.362463] env[62753]: DEBUG oslo_concurrency.lockutils [req-bb1f7144-fe43-4cde-bc29-b6d2c0ab067f req-bcdbad83-2934-49e4-bc56-5d09f22e22ee service nova] Lock "bba5e3cb-c896-4ce7-ac7f-1cf2aba20465-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 968.362788] env[62753]: DEBUG nova.compute.manager [req-bb1f7144-fe43-4cde-bc29-b6d2c0ab067f req-bcdbad83-2934-49e4-bc56-5d09f22e22ee service nova] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] No waiting events found dispatching network-vif-plugged-4832b8bd-7a18-4f8e-a05b-7421417d64ff {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 968.363434] env[62753]: WARNING nova.compute.manager [req-bb1f7144-fe43-4cde-bc29-b6d2c0ab067f req-bcdbad83-2934-49e4-bc56-5d09f22e22ee service nova] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Received unexpected event network-vif-plugged-4832b8bd-7a18-4f8e-a05b-7421417d64ff for instance with vm_state building and task_state spawning. [ 968.455638] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332508, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.080076} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.456059] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 968.456335] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Releasing lock "[datastore1] devstack-image-cache_base/061f0060-350a-4bf7-a725-58acba59de2b/061f0060-350a-4bf7-a725-58acba59de2b.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.456660] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/061f0060-350a-4bf7-a725-58acba59de2b/061f0060-350a-4bf7-a725-58acba59de2b.vmdk to [datastore1] 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5/4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 968.457012] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-38dcf64f-899d-47fc-87a4-edea15c89819 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.464492] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 968.464492] env[62753]: value = "task-1332510" [ 968.464492] env[62753]: _type = "Task" [ 968.464492] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.473578] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332510, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.474638] env[62753]: DEBUG nova.network.neutron [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Successfully updated port: 4832b8bd-7a18-4f8e-a05b-7421417d64ff {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 968.508245] env[62753]: INFO nova.compute.manager [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] instance snapshotting [ 968.509072] env[62753]: DEBUG nova.objects.instance [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lazy-loading 'flavor' on Instance uuid c73b1ae0-dc98-47f7-babf-e96169384785 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 968.510842] env[62753]: DEBUG nova.objects.base [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Object Instance<628239e8-a447-4183-bed9-5550e53fda2b> lazy-loaded attributes: resources,numa_topology {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 968.568565] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332509, 'name': CreateVM_Task, 'duration_secs': 0.332832} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.570089] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Volume attach. Driver type: vmdk {{(pid=62753) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 968.570089] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284643', 'volume_id': '36f91b11-b0aa-4304-aa25-a7ff2fa37218', 'name': 'volume-36f91b11-b0aa-4304-aa25-a7ff2fa37218', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '77cdd901-cc96-4a6f-a696-a65f54a96b1a', 'attached_at': '', 'detached_at': '', 'volume_id': '36f91b11-b0aa-4304-aa25-a7ff2fa37218', 'serial': '36f91b11-b0aa-4304-aa25-a7ff2fa37218'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 968.570412] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 968.571304] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79ae720d-b90a-49e5-b103-aa0a4713212a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.577301] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.577507] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.578089] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 968.579715] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-39fcbf42-0610-4046-9e07-2c5c62415ded {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.598462] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 968.598462] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525df840-45cb-702d-3d3b-3713f3976e01" [ 968.598462] env[62753]: _type = "Task" [ 968.598462] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.602071] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef263e52-2d04-43ba-8ba8-c1ac454b6f51 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.616685] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525df840-45cb-702d-3d3b-3713f3976e01, 'name': SearchDatastore_Task, 'duration_secs': 0.027372} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.638736] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 968.639098] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 968.639423] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.639618] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.639854] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 968.640292] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332507, 'name': ReconfigVM_Task, 'duration_secs': 0.776074} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.648801] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Reconfiguring VM instance instance-00000049 to attach disk [datastore1] volume-36f91b11-b0aa-4304-aa25-a7ff2fa37218/volume-36f91b11-b0aa-4304-aa25-a7ff2fa37218.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 968.650755] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-6c025f7f-d0fd-4352-9253-b3837136e7e9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.652224] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Reconfigured VM instance instance-00000053 to attach disk [datastore2] ff4b295e-2eb2-42dd-b353-08df755f04ea/ff4b295e-2eb2-42dd-b353-08df755f04ea.vmdk or device None with type streamOptimized {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 968.653108] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39702ac2-08de-4a9f-9078-f360b33ca3a7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.666935] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-fac51abc-cb25-4c1f-96ce-511c74c3e716 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.680643] env[62753]: DEBUG oslo_vmware.api [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 968.680643] env[62753]: value = "task-1332512" [ 968.680643] env[62753]: _type = "Task" [ 968.680643] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.680899] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 968.680899] env[62753]: value = "task-1332511" [ 968.680899] env[62753]: _type = "Task" [ 968.680899] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.686419] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 968.686419] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 968.690222] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cc09a0ec-d733-408a-9d02-df870358506d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.696852] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 968.696852] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523bbdea-4e39-6ce5-0094-8c7402de0af8" [ 968.696852] env[62753]: _type = "Task" [ 968.696852] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.703194] env[62753]: DEBUG oslo_vmware.api [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332512, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.703441] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332511, 'name': Rename_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.718061] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523bbdea-4e39-6ce5-0094-8c7402de0af8, 'name': SearchDatastore_Task, 'duration_secs': 0.01268} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 968.723331] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4076bd7b-603d-4844-8496-92b75d6ae159 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.729500] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 968.729500] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52b047f5-13c0-1ef2-fe9c-2fb7d1e499ac" [ 968.729500] env[62753]: _type = "Task" [ 968.729500] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 968.737661] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52b047f5-13c0-1ef2-fe9c-2fb7d1e499ac, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 968.795473] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8527fc59-6bba-40da-82c3-5bedbf9466e0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.803870] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ab362833-b8c7-4a8f-a62b-c1a79accd364 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.839534] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ff86bf-479a-40e8-bb59-62610e3ca2c5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.849264] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18857714-1a21-4b2b-8ab6-f69b8ccd19dc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 968.865519] env[62753]: DEBUG nova.compute.provider_tree [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 968.978504] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "refresh_cache-bba5e3cb-c896-4ce7-ac7f-1cf2aba20465" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 968.978821] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquired lock "refresh_cache-bba5e3cb-c896-4ce7-ac7f-1cf2aba20465" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 968.978821] env[62753]: DEBUG nova.network.neutron [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 968.980329] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332510, 'name': CopyVirtualDisk_Task} progress is 12%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.020661] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38a2ede1-6013-4ae1-869e-a7310afd075a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.045734] env[62753]: DEBUG nova.network.neutron [req-948a5f5d-7d93-47a7-82bd-46793ccb1607 req-d0a9366e-06f9-4bea-9c19-a037f50fdb73 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Updated VIF entry in instance network info cache for port 0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 969.046225] env[62753]: DEBUG nova.network.neutron [req-948a5f5d-7d93-47a7-82bd-46793ccb1607 req-d0a9366e-06f9-4bea-9c19-a037f50fdb73 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Updating instance_info_cache with network_info: [{"id": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "address": "fa:16:3e:6b:34:1e", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cd05bfa-82", "ovs_interfaceid": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.049106] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d195118f-4f2e-45f8-a5b0-2743d8163638 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.208277] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332511, 'name': Rename_Task, 'duration_secs': 0.224385} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.211971] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 969.212331] env[62753]: DEBUG oslo_vmware.api [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332512, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.212574] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-240d17cf-d508-40bd-bef6-39ea51b72bb5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.222908] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 969.222908] env[62753]: value = "task-1332513" [ 969.222908] env[62753]: _type = "Task" [ 969.222908] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.237813] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332513, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.244659] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52b047f5-13c0-1ef2-fe9c-2fb7d1e499ac, 'name': SearchDatastore_Task, 'duration_secs': 0.017591} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 969.244900] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.245228] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 53a76eb0-8b0c-4e47-8387-75df4c0b191f/53a76eb0-8b0c-4e47-8387-75df4c0b191f.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 969.245549] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-88cc5bb0-3de8-44ba-aa7c-e0d22ed5a132 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.253538] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 969.253538] env[62753]: value = "task-1332514" [ 969.253538] env[62753]: _type = "Task" [ 969.253538] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.264418] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332514, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.370867] env[62753]: DEBUG nova.scheduler.client.report [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 969.477278] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332510, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.527412] env[62753]: DEBUG nova.network.neutron [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 969.548893] env[62753]: DEBUG oslo_concurrency.lockutils [req-948a5f5d-7d93-47a7-82bd-46793ccb1607 req-d0a9366e-06f9-4bea-9c19-a037f50fdb73 service nova] Releasing lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 969.564715] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Creating Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 969.565170] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-1692dbc2-4e60-4bc2-a0a4-9ec82db2bc80 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 969.574333] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 969.574333] env[62753]: value = "task-1332515" [ 969.574333] env[62753]: _type = "Task" [ 969.574333] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 969.584082] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332515, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.661961] env[62753]: DEBUG nova.network.neutron [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Updating instance_info_cache with network_info: [{"id": "4832b8bd-7a18-4f8e-a05b-7421417d64ff", "address": "fa:16:3e:a5:4c:97", "network": {"id": "22a90636-182f-4944-893f-329f446b07c4", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1513076888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "37368e15fa5a4a0a97ac1c22fcf0bb80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4832b8bd-7a", "ovs_interfaceid": "4832b8bd-7a18-4f8e-a05b-7421417d64ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 969.704847] env[62753]: DEBUG oslo_vmware.api [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332512, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.735597] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332513, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.764545] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332514, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 969.877041] env[62753]: DEBUG oslo_concurrency.lockutils [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.372s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 969.880142] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.774s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 969.880577] env[62753]: DEBUG nova.objects.instance [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lazy-loading 'resources' on Instance uuid abd5ec6e-7b12-433b-9009-f9108ab4e318 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 969.978053] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332510, 'name': CopyVirtualDisk_Task} progress is 60%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.088156] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332515, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.165501] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Releasing lock "refresh_cache-bba5e3cb-c896-4ce7-ac7f-1cf2aba20465" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 970.166074] env[62753]: DEBUG nova.compute.manager [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Instance network_info: |[{"id": "4832b8bd-7a18-4f8e-a05b-7421417d64ff", "address": "fa:16:3e:a5:4c:97", "network": {"id": "22a90636-182f-4944-893f-329f446b07c4", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1513076888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "37368e15fa5a4a0a97ac1c22fcf0bb80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4832b8bd-7a", "ovs_interfaceid": "4832b8bd-7a18-4f8e-a05b-7421417d64ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 970.166812] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a5:4c:97', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8e7f6f41-f4eb-4832-a390-730fca1cf717', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '4832b8bd-7a18-4f8e-a05b-7421417d64ff', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 970.180301] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Creating folder: Project (37368e15fa5a4a0a97ac1c22fcf0bb80). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 970.180808] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-9713a79f-6700-421a-b0fb-f9e03c7bcb7c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.207972] env[62753]: DEBUG oslo_vmware.api [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332512, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.212662] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Created folder: Project (37368e15fa5a4a0a97ac1c22fcf0bb80) in parent group-v284541. [ 970.212818] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Creating folder: Instances. Parent ref: group-v284645. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 970.213022] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-886bfe4c-b082-4c7b-96cd-445ffb1a5b06 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.229767] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Created folder: Instances in parent group-v284645. [ 970.229968] env[62753]: DEBUG oslo.service.loopingcall [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.230667] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 970.230916] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cd3ca6cc-021d-4213-99f6-526eede800fb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.253205] env[62753]: DEBUG oslo_vmware.api [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332513, 'name': PowerOnVM_Task, 'duration_secs': 0.700982} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.254014] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 970.254226] env[62753]: INFO nova.compute.manager [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Took 18.34 seconds to spawn the instance on the hypervisor. [ 970.254411] env[62753]: DEBUG nova.compute.manager [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 970.255375] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e211a76-18f5-47d0-b591-aa86bf9d5033 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.264094] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 970.264094] env[62753]: value = "task-1332518" [ 970.264094] env[62753]: _type = "Task" [ 970.264094] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.277838] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332514, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.283427] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332518, 'name': CreateVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.392178] env[62753]: DEBUG oslo_concurrency.lockutils [None req-425d7b69-93eb-41c4-ae5f-7070f4f8aa3c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "628239e8-a447-4183-bed9-5550e53fda2b" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 25.295s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.394123] env[62753]: DEBUG oslo_concurrency.lockutils [None req-de351a37-b138-4888-adf5-f77f8aa16990 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "628239e8-a447-4183-bed9-5550e53fda2b" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 3.898s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.394540] env[62753]: DEBUG oslo_concurrency.lockutils [None req-de351a37-b138-4888-adf5-f77f8aa16990 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "628239e8-a447-4183-bed9-5550e53fda2b-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 970.394833] env[62753]: DEBUG oslo_concurrency.lockutils [None req-de351a37-b138-4888-adf5-f77f8aa16990 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "628239e8-a447-4183-bed9-5550e53fda2b-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 970.395152] env[62753]: DEBUG oslo_concurrency.lockutils [None req-de351a37-b138-4888-adf5-f77f8aa16990 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "628239e8-a447-4183-bed9-5550e53fda2b-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 970.402020] env[62753]: DEBUG nova.compute.manager [req-89a754a6-91ab-4a2e-806d-87417355e90d req-4ce8c08b-2ae7-412c-b3b7-a6b8d1fbd0e1 service nova] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Received event network-changed-4832b8bd-7a18-4f8e-a05b-7421417d64ff {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 970.402629] env[62753]: DEBUG nova.compute.manager [req-89a754a6-91ab-4a2e-806d-87417355e90d req-4ce8c08b-2ae7-412c-b3b7-a6b8d1fbd0e1 service nova] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Refreshing instance network info cache due to event network-changed-4832b8bd-7a18-4f8e-a05b-7421417d64ff. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 970.403157] env[62753]: DEBUG oslo_concurrency.lockutils [req-89a754a6-91ab-4a2e-806d-87417355e90d req-4ce8c08b-2ae7-412c-b3b7-a6b8d1fbd0e1 service nova] Acquiring lock "refresh_cache-bba5e3cb-c896-4ce7-ac7f-1cf2aba20465" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 970.403157] env[62753]: DEBUG oslo_concurrency.lockutils [req-89a754a6-91ab-4a2e-806d-87417355e90d req-4ce8c08b-2ae7-412c-b3b7-a6b8d1fbd0e1 service nova] Acquired lock "refresh_cache-bba5e3cb-c896-4ce7-ac7f-1cf2aba20465" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 970.403428] env[62753]: DEBUG nova.network.neutron [req-89a754a6-91ab-4a2e-806d-87417355e90d req-4ce8c08b-2ae7-412c-b3b7-a6b8d1fbd0e1 service nova] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Refreshing network info cache for port 4832b8bd-7a18-4f8e-a05b-7421417d64ff {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 970.405520] env[62753]: INFO nova.compute.manager [None req-de351a37-b138-4888-adf5-f77f8aa16990 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Terminating instance [ 970.408883] env[62753]: DEBUG nova.compute.manager [None req-de351a37-b138-4888-adf5-f77f8aa16990 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 970.409150] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-de351a37-b138-4888-adf5-f77f8aa16990 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 970.409523] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-772ae185-02a9-4385-949b-1be38102d3b3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.425486] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-debd10ec-0ec3-4f26-b8cb-e458e782e70f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.468021] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-de351a37-b138-4888-adf5-f77f8aa16990 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 628239e8-a447-4183-bed9-5550e53fda2b could not be found. [ 970.468310] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-de351a37-b138-4888-adf5-f77f8aa16990 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 970.468556] env[62753]: INFO nova.compute.manager [None req-de351a37-b138-4888-adf5-f77f8aa16990 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Took 0.06 seconds to destroy the instance on the hypervisor. [ 970.469007] env[62753]: DEBUG oslo.service.loopingcall [None req-de351a37-b138-4888-adf5-f77f8aa16990 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 970.476900] env[62753]: DEBUG nova.compute.manager [-] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 970.477032] env[62753]: DEBUG nova.network.neutron [-] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 970.487586] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332510, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.593928] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332515, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.684957] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd117697-c950-4337-a1a8-04feb5831b61 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.694624] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7cb85d62-a4a9-4783-8b7e-718129b99d18 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.737921] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e8a3557-b78e-444d-8690-ad7aa2722609 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.740924] env[62753]: DEBUG oslo_vmware.api [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332512, 'name': ReconfigVM_Task, 'duration_secs': 1.665199} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.741354] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Reconfigured VM instance instance-00000049 to attach disk [datastore1] volume-36f91b11-b0aa-4304-aa25-a7ff2fa37218/volume-36f91b11-b0aa-4304-aa25-a7ff2fa37218.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 970.746902] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9e3cace7-1873-43d1-9bd5-2bbdeef16c91 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.765049] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a4f4677-5808-48ec-9d36-0278be3bd256 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 970.772538] env[62753]: DEBUG oslo_vmware.api [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 970.772538] env[62753]: value = "task-1332519" [ 970.772538] env[62753]: _type = "Task" [ 970.772538] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 970.789357] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332514, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.808884] env[62753]: DEBUG nova.compute.provider_tree [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 970.815023] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332518, 'name': CreateVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.815023] env[62753]: INFO nova.compute.manager [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Took 28.48 seconds to build instance. [ 970.817792] env[62753]: DEBUG oslo_vmware.api [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332519, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 970.980643] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332510, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.452965} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 970.981879] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/061f0060-350a-4bf7-a725-58acba59de2b/061f0060-350a-4bf7-a725-58acba59de2b.vmdk to [datastore1] 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5/4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 970.982857] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7827e536-bce6-4aa1-867d-1658d46a09a5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.011056] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Reconfiguring VM instance instance-0000003e to attach disk [datastore1] 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5/4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5.vmdk or device None with type streamOptimized {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 971.015138] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8aa12def-c3c7-4eda-82fd-dae36a3bf2a6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.042812] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 971.042812] env[62753]: value = "task-1332520" [ 971.042812] env[62753]: _type = "Task" [ 971.042812] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.055994] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332520, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.088942] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "ff4b295e-2eb2-42dd-b353-08df755f04ea" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.089302] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332515, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.188175] env[62753]: DEBUG nova.network.neutron [req-89a754a6-91ab-4a2e-806d-87417355e90d req-4ce8c08b-2ae7-412c-b3b7-a6b8d1fbd0e1 service nova] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Updated VIF entry in instance network info cache for port 4832b8bd-7a18-4f8e-a05b-7421417d64ff. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 971.188710] env[62753]: DEBUG nova.network.neutron [req-89a754a6-91ab-4a2e-806d-87417355e90d req-4ce8c08b-2ae7-412c-b3b7-a6b8d1fbd0e1 service nova] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Updating instance_info_cache with network_info: [{"id": "4832b8bd-7a18-4f8e-a05b-7421417d64ff", "address": "fa:16:3e:a5:4c:97", "network": {"id": "22a90636-182f-4944-893f-329f446b07c4", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1513076888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "37368e15fa5a4a0a97ac1c22fcf0bb80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4832b8bd-7a", "ovs_interfaceid": "4832b8bd-7a18-4f8e-a05b-7421417d64ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.271178] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332514, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.991741} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.271504] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 53a76eb0-8b0c-4e47-8387-75df4c0b191f/53a76eb0-8b0c-4e47-8387-75df4c0b191f.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 971.271799] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 971.272123] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-04748ff3-f0d4-4b2c-983a-dab84d5aa67f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.282920] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 971.282920] env[62753]: value = "task-1332521" [ 971.282920] env[62753]: _type = "Task" [ 971.282920] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.290579] env[62753]: DEBUG oslo_vmware.api [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332519, 'name': ReconfigVM_Task, 'duration_secs': 0.193034} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.290804] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332518, 'name': CreateVM_Task, 'duration_secs': 0.616387} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.294034] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284643', 'volume_id': '36f91b11-b0aa-4304-aa25-a7ff2fa37218', 'name': 'volume-36f91b11-b0aa-4304-aa25-a7ff2fa37218', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '77cdd901-cc96-4a6f-a696-a65f54a96b1a', 'attached_at': '', 'detached_at': '', 'volume_id': '36f91b11-b0aa-4304-aa25-a7ff2fa37218', 'serial': '36f91b11-b0aa-4304-aa25-a7ff2fa37218'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 971.295597] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 971.297643] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.298362] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.298416] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 971.299412] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6bb2bca-fd9c-4f59-b297-bcbc3b3269e8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.305808] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332521, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.307776] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 971.307776] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5205610b-e35a-c6eb-d423-ec0a15759cac" [ 971.307776] env[62753]: _type = "Task" [ 971.307776] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.313395] env[62753]: DEBUG nova.scheduler.client.report [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 971.318045] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7ae3599b-90b3-42f0-bba8-946bb858622d tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "ff4b295e-2eb2-42dd-b353-08df755f04ea" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 30.615s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.318785] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "ff4b295e-2eb2-42dd-b353-08df755f04ea" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.230s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.319055] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "ff4b295e-2eb2-42dd-b353-08df755f04ea-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 971.319294] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "ff4b295e-2eb2-42dd-b353-08df755f04ea-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.319507] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "ff4b295e-2eb2-42dd-b353-08df755f04ea-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.325996] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5205610b-e35a-c6eb-d423-ec0a15759cac, 'name': SearchDatastore_Task, 'duration_secs': 0.015508} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.326657] env[62753]: INFO nova.compute.manager [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Terminating instance [ 971.328386] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.328748] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 971.329067] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 971.329350] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 971.329595] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 971.330416] env[62753]: DEBUG nova.compute.manager [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 971.330735] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 971.331886] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-a7fadb1e-2079-4617-af18-9ae57b23ad18 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.335519] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abb7ad99-7ef7-462a-b88a-bbf95d936eee {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.344821] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 971.345144] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-7865171f-76dd-4f57-b0ae-aad41cfd9449 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.348173] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 971.348374] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 971.349450] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3e3ad26-d650-4dab-ab12-0e89df3f65f0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.355473] env[62753]: DEBUG oslo_vmware.api [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 971.355473] env[62753]: value = "task-1332522" [ 971.355473] env[62753]: _type = "Task" [ 971.355473] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.357363] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 971.357363] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]526b423f-e5c2-8c0b-5b31-70c4b679e2a2" [ 971.357363] env[62753]: _type = "Task" [ 971.357363] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.371887] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]526b423f-e5c2-8c0b-5b31-70c4b679e2a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.375681] env[62753]: DEBUG oslo_vmware.api [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332522, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.468870] env[62753]: DEBUG nova.network.neutron [-] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 971.557747] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332520, 'name': ReconfigVM_Task, 'duration_secs': 0.338208} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.557747] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Reconfigured VM instance instance-0000003e to attach disk [datastore1] 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5/4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5.vmdk or device None with type streamOptimized {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 971.557747] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-f52e4c73-6b9c-43a7-b15e-bdcdf8c678b6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.564096] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 971.564096] env[62753]: value = "task-1332523" [ 971.564096] env[62753]: _type = "Task" [ 971.564096] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.573798] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332523, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.590013] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332515, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.692707] env[62753]: DEBUG oslo_concurrency.lockutils [req-89a754a6-91ab-4a2e-806d-87417355e90d req-4ce8c08b-2ae7-412c-b3b7-a6b8d1fbd0e1 service nova] Releasing lock "refresh_cache-bba5e3cb-c896-4ce7-ac7f-1cf2aba20465" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 971.805865] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332521, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.078737} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.806409] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 971.807242] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8ab76b6-670e-4ad9-a686-0b5026c1a35c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.832225] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.951s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 971.850036] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Reconfiguring VM instance instance-00000054 to attach disk [datastore1] 53a76eb0-8b0c-4e47-8387-75df4c0b191f/53a76eb0-8b0c-4e47-8387-75df4c0b191f.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 971.855559] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.273s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 971.855783] env[62753]: INFO nova.compute.claims [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 971.858874] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2cd3df35-5ed6-4bd5-86c6-969f7a3c2a1d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.894626] env[62753]: DEBUG oslo_vmware.api [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332522, 'name': PowerOffVM_Task, 'duration_secs': 0.219974} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.895404] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]526b423f-e5c2-8c0b-5b31-70c4b679e2a2, 'name': SearchDatastore_Task, 'duration_secs': 0.014305} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 971.897011] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 971.897260] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 971.898211] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 971.898211] env[62753]: value = "task-1332524" [ 971.898211] env[62753]: _type = "Task" [ 971.898211] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.898442] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a9e3a1da-5860-4cfd-9216-61bbbec684d7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.900135] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-90fe769f-fc7b-45b0-97cf-abeb62817db5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 971.903730] env[62753]: INFO nova.scheduler.client.report [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Deleted allocations for instance abd5ec6e-7b12-433b-9009-f9108ab4e318 [ 971.918617] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332524, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.920326] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 971.920326] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]521c1daf-bc72-f349-47ef-42ed6d8c102d" [ 971.920326] env[62753]: _type = "Task" [ 971.920326] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 971.931060] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]521c1daf-bc72-f349-47ef-42ed6d8c102d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 971.979033] env[62753]: INFO nova.compute.manager [-] [instance: 628239e8-a447-4183-bed9-5550e53fda2b] Took 1.50 seconds to deallocate network for instance. [ 971.992248] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 971.992248] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 971.992369] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Deleting the datastore file [datastore2] ff4b295e-2eb2-42dd-b353-08df755f04ea {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 971.992574] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6f6969df-9de8-445e-bd28-e59bc741af24 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.000546] env[62753]: DEBUG oslo_vmware.api [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 972.000546] env[62753]: value = "task-1332526" [ 972.000546] env[62753]: _type = "Task" [ 972.000546] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.010086] env[62753]: DEBUG oslo_vmware.api [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332526, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.079086] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332523, 'name': Rename_Task, 'duration_secs': 0.151698} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.082634] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 972.082959] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3d8c424f-0db3-41ef-905a-5f7adb804ab8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.091501] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332515, 'name': CreateSnapshot_Task, 'duration_secs': 2.1407} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.092967] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Created Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 972.093353] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 972.093353] env[62753]: value = "task-1332527" [ 972.093353] env[62753]: _type = "Task" [ 972.093353] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.094149] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b3890d3-8515-4406-9831-76da976e64e7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.106538] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332527, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.377562] env[62753]: DEBUG nova.objects.instance [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lazy-loading 'flavor' on Instance uuid 77cdd901-cc96-4a6f-a696-a65f54a96b1a {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 972.414903] env[62753]: DEBUG oslo_concurrency.lockutils [None req-dcfa1a18-9dfb-412e-ab76-57f359fd2cd1 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "abd5ec6e-7b12-433b-9009-f9108ab4e318" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.078s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.423669] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332524, 'name': ReconfigVM_Task, 'duration_secs': 0.313745} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.428174] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Reconfigured VM instance instance-00000054 to attach disk [datastore1] 53a76eb0-8b0c-4e47-8387-75df4c0b191f/53a76eb0-8b0c-4e47-8387-75df4c0b191f.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 972.429358] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cb1df833-2c50-4af6-bf6f-5d5a0bbc08f8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.439944] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]521c1daf-bc72-f349-47ef-42ed6d8c102d, 'name': SearchDatastore_Task, 'duration_secs': 0.021765} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.441692] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 972.441984] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] bba5e3cb-c896-4ce7-ac7f-1cf2aba20465/bba5e3cb-c896-4ce7-ac7f-1cf2aba20465.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 972.442363] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 972.442363] env[62753]: value = "task-1332528" [ 972.442363] env[62753]: _type = "Task" [ 972.442363] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.442753] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-67288741-fa22-4ec5-9b86-f798d83685bc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.457154] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332528, 'name': Rename_Task} progress is 10%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.459704] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 972.459704] env[62753]: value = "task-1332529" [ 972.459704] env[62753]: _type = "Task" [ 972.459704] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.470479] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332529, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.517205] env[62753]: DEBUG oslo_vmware.api [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332526, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.242678} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.517495] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 972.517720] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 972.517920] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 972.518123] env[62753]: INFO nova.compute.manager [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Took 1.19 seconds to destroy the instance on the hypervisor. [ 972.518393] env[62753]: DEBUG oslo.service.loopingcall [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 972.518627] env[62753]: DEBUG nova.compute.manager [-] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 972.518727] env[62753]: DEBUG nova.network.neutron [-] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 972.610602] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332527, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.621240] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Creating linked-clone VM from snapshot {{(pid=62753) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 972.622089] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-fbd47b84-d01f-4a89-8015-3bbce5e1d8d7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.632019] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 972.632019] env[62753]: value = "task-1332530" [ 972.632019] env[62753]: _type = "Task" [ 972.632019] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.646041] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332530, 'name': CloneVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.648057] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "8066aeb2-7be4-46e3-aba6-124fef342a49" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.648533] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "8066aeb2-7be4-46e3-aba6-124fef342a49" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.648719] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "8066aeb2-7be4-46e3-aba6-124fef342a49-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 972.649077] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "8066aeb2-7be4-46e3-aba6-124fef342a49-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 972.649358] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "8066aeb2-7be4-46e3-aba6-124fef342a49-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.655825] env[62753]: INFO nova.compute.manager [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Terminating instance [ 972.661179] env[62753]: DEBUG nova.compute.manager [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 972.661456] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 972.662570] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-881a19a7-ee89-48ad-9878-7088d42f9a39 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.674161] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 972.674748] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-04e1bbcc-18e6-4c09-b4bb-72b970c65296 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.687741] env[62753]: DEBUG oslo_vmware.api [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 972.687741] env[62753]: value = "task-1332531" [ 972.687741] env[62753]: _type = "Task" [ 972.687741] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.699679] env[62753]: DEBUG oslo_vmware.api [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332531, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.889423] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3477cb5c-3874-4165-b362-a07bf3435cdb tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 8.938s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 972.958763] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332528, 'name': Rename_Task, 'duration_secs': 0.227922} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 972.962207] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 972.962957] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c2f4a311-804a-4f28-acba-bd111d2cdefa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 972.977423] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332529, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 972.979906] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 972.979906] env[62753]: value = "task-1332532" [ 972.979906] env[62753]: _type = "Task" [ 972.979906] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 972.993206] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332532, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.013997] env[62753]: DEBUG oslo_concurrency.lockutils [None req-de351a37-b138-4888-adf5-f77f8aa16990 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "628239e8-a447-4183-bed9-5550e53fda2b" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 2.619s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.112304] env[62753]: DEBUG oslo_vmware.api [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332527, 'name': PowerOnVM_Task, 'duration_secs': 0.625597} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.112700] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 973.150039] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332530, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.167013] env[62753]: DEBUG nova.compute.manager [req-616d94fc-4db1-43bf-9855-d5c163321ec8 req-fd11d257-cab9-4128-b78d-09e1d6afed80 service nova] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Received event network-vif-deleted-2003ddeb-12c2-4450-a45b-cc180b041d34 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 973.167693] env[62753]: INFO nova.compute.manager [req-616d94fc-4db1-43bf-9855-d5c163321ec8 req-fd11d257-cab9-4128-b78d-09e1d6afed80 service nova] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Neutron deleted interface 2003ddeb-12c2-4450-a45b-cc180b041d34; detaching it from the instance and deleting it from the info cache [ 973.167777] env[62753]: DEBUG nova.network.neutron [req-616d94fc-4db1-43bf-9855-d5c163321ec8 req-fd11d257-cab9-4128-b78d-09e1d6afed80 service nova] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.172127] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b4ac43da-7220-4f2c-9bcc-64d00fbb5714 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.185942] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2e5b3fe8-8052-47fc-a398-0610c4b70381 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.233383] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-828fa1a3-0522-4b50-b847-d9045a714efb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.237451] env[62753]: DEBUG oslo_vmware.api [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332531, 'name': PowerOffVM_Task, 'duration_secs': 0.46736} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.237451] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 973.237451] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 973.237713] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a2ccbdfb-3943-4974-9fe9-9372b1d2984f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.244281] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bafa1a2f-d263-47d4-b987-0a272f55f87a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.264198] env[62753]: DEBUG nova.compute.provider_tree [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 973.266898] env[62753]: DEBUG nova.compute.manager [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 973.267930] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa25fbc3-42f7-495e-aaf9-c4903a7c7e4d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.354800] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 973.355149] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 973.355379] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Deleting the datastore file [datastore2] 8066aeb2-7be4-46e3-aba6-124fef342a49 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 973.355667] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ae8b826a-9f53-43a0-b995-f7d9d7453da8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.362967] env[62753]: DEBUG oslo_vmware.api [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 973.362967] env[62753]: value = "task-1332534" [ 973.362967] env[62753]: _type = "Task" [ 973.362967] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.373048] env[62753]: DEBUG oslo_vmware.api [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332534, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.475313] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332529, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.634834} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.476396] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] bba5e3cb-c896-4ce7-ac7f-1cf2aba20465/bba5e3cb-c896-4ce7-ac7f-1cf2aba20465.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 973.476396] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 973.477092] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2a0a392d-c900-4b44-a51e-8b8bfd5b11a2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.486064] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 973.486064] env[62753]: value = "task-1332535" [ 973.486064] env[62753]: _type = "Task" [ 973.486064] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 973.494495] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332532, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.500748] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332535, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.574010] env[62753]: DEBUG nova.network.neutron [-] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 973.646759] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332530, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 973.681021] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-44f0501b-bcbf-4a56-a4e8-a98a4f2eed9b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.689279] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5deebac-1b53-4318-acd0-61a9b8bfbcdc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 973.731073] env[62753]: DEBUG nova.compute.manager [req-616d94fc-4db1-43bf-9855-d5c163321ec8 req-fd11d257-cab9-4128-b78d-09e1d6afed80 service nova] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Detach interface failed, port_id=2003ddeb-12c2-4450-a45b-cc180b041d34, reason: Instance ff4b295e-2eb2-42dd-b353-08df755f04ea could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 973.779633] env[62753]: DEBUG nova.scheduler.client.report [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 973.790126] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c865fad-5c2f-49ec-ba11-dadab03b2c7a tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 27.979s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 973.880038] env[62753]: DEBUG oslo_vmware.api [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332534, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.321639} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.880038] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 973.880038] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 973.880038] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 973.880038] env[62753]: INFO nova.compute.manager [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Took 1.22 seconds to destroy the instance on the hypervisor. [ 973.880038] env[62753]: DEBUG oslo.service.loopingcall [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 973.880038] env[62753]: DEBUG nova.compute.manager [-] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 973.880038] env[62753]: DEBUG nova.network.neutron [-] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 973.901502] env[62753]: DEBUG oslo_concurrency.lockutils [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 973.901879] env[62753]: DEBUG oslo_concurrency.lockutils [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 973.996263] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332535, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087958} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 973.999819] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 974.000457] env[62753]: DEBUG oslo_vmware.api [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332532, 'name': PowerOnVM_Task, 'duration_secs': 0.791369} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 974.004138] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5055e80d-e0c3-4cfc-9766-8c782a6e6a53 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.004895] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 974.004895] env[62753]: INFO nova.compute.manager [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Took 8.27 seconds to spawn the instance on the hypervisor. [ 974.004895] env[62753]: DEBUG nova.compute.manager [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 974.005804] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-959d155d-9c68-4c6b-aabe-a343e74d6dea {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.040504] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] bba5e3cb-c896-4ce7-ac7f-1cf2aba20465/bba5e3cb-c896-4ce7-ac7f-1cf2aba20465.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 974.041378] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2224a465-9405-49e7-a337-372adc41bdbd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.064609] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 974.064609] env[62753]: value = "task-1332536" [ 974.064609] env[62753]: _type = "Task" [ 974.064609] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.075594] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332536, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.076155] env[62753]: INFO nova.compute.manager [-] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Took 1.56 seconds to deallocate network for instance. [ 974.150910] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332530, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.286224] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.432s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.286896] env[62753]: DEBUG nova.compute.manager [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 974.290559] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.398s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.290877] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.327764] env[62753]: INFO nova.scheduler.client.report [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Deleted allocations for instance a401909c-5978-492b-a4fa-b17d38a5af44 [ 974.340983] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "91fc519e-c9bb-4605-bdc6-4f463561b988" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.340983] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "91fc519e-c9bb-4605-bdc6-4f463561b988" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.405514] env[62753]: INFO nova.compute.manager [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Detaching volume 368def30-47d9-476e-8669-d2beafd3fba1 [ 974.447862] env[62753]: INFO nova.virt.block_device [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Attempting to driver detach volume 368def30-47d9-476e-8669-d2beafd3fba1 from mountpoint /dev/sdb [ 974.447862] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Volume detach. Driver type: vmdk {{(pid=62753) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 974.447862] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284636', 'volume_id': '368def30-47d9-476e-8669-d2beafd3fba1', 'name': 'volume-368def30-47d9-476e-8669-d2beafd3fba1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '77cdd901-cc96-4a6f-a696-a65f54a96b1a', 'attached_at': '', 'detached_at': '', 'volume_id': '368def30-47d9-476e-8669-d2beafd3fba1', 'serial': '368def30-47d9-476e-8669-d2beafd3fba1'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 974.449030] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c23eed29-f9a5-49a0-a02d-54e817c08d88 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.478517] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f72b7c2-4ff2-45af-8b5a-209f19880106 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.491033] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ec67d45-5769-41f0-ae0c-fb1dc7200689 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.522116] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4294ac8f-17ec-4406-b6e7-9a193230d256 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.539485] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] The volume has not been displaced from its original location: [datastore2] volume-368def30-47d9-476e-8669-d2beafd3fba1/volume-368def30-47d9-476e-8669-d2beafd3fba1.vmdk. No consolidation needed. {{(pid=62753) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 974.545169] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Reconfiguring VM instance instance-00000049 to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 974.549977] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-584f58df-ebc2-4893-89bb-6681df308158 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 974.574343] env[62753]: INFO nova.compute.manager [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Took 13.34 seconds to build instance. [ 974.577385] env[62753]: DEBUG oslo_vmware.api [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 974.577385] env[62753]: value = "task-1332537" [ 974.577385] env[62753]: _type = "Task" [ 974.577385] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 974.582726] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 974.583042] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 974.583287] env[62753]: DEBUG nova.objects.instance [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lazy-loading 'resources' on Instance uuid ff4b295e-2eb2-42dd-b353-08df755f04ea {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 974.588628] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332536, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.595546] env[62753]: DEBUG oslo_vmware.api [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332537, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.645213] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332530, 'name': CloneVM_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 974.670708] env[62753]: DEBUG nova.network.neutron [-] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 974.796864] env[62753]: DEBUG nova.compute.utils [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 974.799194] env[62753]: DEBUG nova.compute.manager [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 974.799194] env[62753]: DEBUG nova.network.neutron [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 974.837608] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62cfde5c-fd40-4795-a24f-74442084b62f tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "a401909c-5978-492b-a4fa-b17d38a5af44" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.789s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 974.844074] env[62753]: DEBUG nova.compute.manager [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 974.862234] env[62753]: DEBUG nova.policy [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b6207815e8041b4bba3069635d7d962', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6c8a6bc946a469fa85cc3ab80d7333a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 975.198981] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332536, 'name': ReconfigVM_Task, 'duration_secs': 0.595123} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.198981] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3e5a8219-d9f6-4abc-90d6-3a5969b5a9d5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.884s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 975.198981] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Reconfigured VM instance instance-00000055 to attach disk [datastore1] bba5e3cb-c896-4ce7-ac7f-1cf2aba20465/bba5e3cb-c896-4ce7-ac7f-1cf2aba20465.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 975.198981] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-57b10d53-ec41-46c2-a974-14d864b4c3a9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.198981] env[62753]: DEBUG oslo_vmware.api [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332537, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.198981] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 975.198981] env[62753]: value = "task-1332538" [ 975.198981] env[62753]: _type = "Task" [ 975.198981] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 975.198981] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332538, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.198981] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332530, 'name': CloneVM_Task, 'duration_secs': 2.050283} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 975.198981] env[62753]: INFO nova.virt.vmwareapi.vmops [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Created linked-clone VM from snapshot [ 975.198981] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b392f76-00f5-4607-8431-fc09ca88cacc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.198981] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Uploading image 3ebd5a0c-fc07-4177-9615-bef143e04e30 {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 975.198981] env[62753]: INFO nova.compute.manager [-] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Took 1.30 seconds to deallocate network for instance. [ 975.212905] env[62753]: DEBUG nova.compute.manager [req-c866e98f-228a-4c89-97c1-4b1d6d394637 req-f916e8f5-c784-4aa2-8d78-5fa9d9071103 service nova] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Received event network-vif-deleted-f433dba2-2032-4957-8ba6-06568d931ccc {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 975.216518] env[62753]: DEBUG oslo_vmware.rw_handles [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 975.216518] env[62753]: value = "vm-284649" [ 975.216518] env[62753]: _type = "VirtualMachine" [ 975.216518] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 975.217206] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-bdbc9e63-bb6e-4133-a4e0-414437dcafa9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.231214] env[62753]: DEBUG oslo_vmware.rw_handles [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lease: (returnval){ [ 975.231214] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5293d8d7-80a5-bbf7-68b1-e31d7245df4c" [ 975.231214] env[62753]: _type = "HttpNfcLease" [ 975.231214] env[62753]: } obtained for exporting VM: (result){ [ 975.231214] env[62753]: value = "vm-284649" [ 975.231214] env[62753]: _type = "VirtualMachine" [ 975.231214] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 975.231590] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the lease: (returnval){ [ 975.231590] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5293d8d7-80a5-bbf7-68b1-e31d7245df4c" [ 975.231590] env[62753]: _type = "HttpNfcLease" [ 975.231590] env[62753]: } to be ready. {{(pid=62753) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 975.245175] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 975.245175] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5293d8d7-80a5-bbf7-68b1-e31d7245df4c" [ 975.245175] env[62753]: _type = "HttpNfcLease" [ 975.245175] env[62753]: } is ready. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 975.245551] env[62753]: DEBUG oslo_vmware.rw_handles [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 975.245551] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5293d8d7-80a5-bbf7-68b1-e31d7245df4c" [ 975.245551] env[62753]: _type = "HttpNfcLease" [ 975.245551] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 975.246433] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdb25bff-3093-4a07-ba82-32ec6b4a27e8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.258510] env[62753]: DEBUG oslo_vmware.rw_handles [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e209be-2ca4-dc01-c311-ed193decc7f3/disk-0.vmdk from lease info. {{(pid=62753) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 975.258787] env[62753]: DEBUG oslo_vmware.rw_handles [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e209be-2ca4-dc01-c311-ed193decc7f3/disk-0.vmdk for reading. {{(pid=62753) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 975.332203] env[62753]: DEBUG nova.compute.manager [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 975.339762] env[62753]: DEBUG nova.network.neutron [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Successfully created port: c86749ca-c0b7-4c20-a32a-fed94bb47e4d {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 975.375256] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.380846] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-38055845-56d3-4a35-946e-bf815e33d6f5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.437802] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8a1d6cd-b6a3-423b-bb92-92f7f209dedd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.448309] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-237084b2-d43b-445a-b199-512fd331befa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.481496] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38b2e91f-3a6c-42f1-9995-fee8649b1b07 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.490659] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73a2c1b5-8370-49a2-849b-5e84191c46e6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 975.507401] env[62753]: DEBUG nova.compute.provider_tree [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 975.592581] env[62753]: DEBUG oslo_vmware.api [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332537, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.621157] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332538, 'name': Rename_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 975.685522] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 975.951963] env[62753]: DEBUG nova.compute.manager [req-483bd152-6b17-4a3b-9ebd-a2dfa9ec67d7 req-ddd63173-0d95-4962-9aa0-7a9f9d748d0a service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Received event network-changed-12ca481a-0cb3-465e-b04c-802ae3ce6438 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 975.952402] env[62753]: DEBUG nova.compute.manager [req-483bd152-6b17-4a3b-9ebd-a2dfa9ec67d7 req-ddd63173-0d95-4962-9aa0-7a9f9d748d0a service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Refreshing instance network info cache due to event network-changed-12ca481a-0cb3-465e-b04c-802ae3ce6438. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 975.952402] env[62753]: DEBUG oslo_concurrency.lockutils [req-483bd152-6b17-4a3b-9ebd-a2dfa9ec67d7 req-ddd63173-0d95-4962-9aa0-7a9f9d748d0a service nova] Acquiring lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 975.952559] env[62753]: DEBUG oslo_concurrency.lockutils [req-483bd152-6b17-4a3b-9ebd-a2dfa9ec67d7 req-ddd63173-0d95-4962-9aa0-7a9f9d748d0a service nova] Acquired lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 975.952714] env[62753]: DEBUG nova.network.neutron [req-483bd152-6b17-4a3b-9ebd-a2dfa9ec67d7 req-ddd63173-0d95-4962-9aa0-7a9f9d748d0a service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Refreshing network info cache for port 12ca481a-0cb3-465e-b04c-802ae3ce6438 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 976.012389] env[62753]: DEBUG nova.scheduler.client.report [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 976.096251] env[62753]: DEBUG oslo_vmware.api [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332537, 'name': ReconfigVM_Task, 'duration_secs': 1.378105} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.096251] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Reconfigured VM instance instance-00000049 to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 976.106396] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4f808770-7afe-4e9d-a0fb-4b2e1f1d19b4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.129279] env[62753]: DEBUG oslo_vmware.api [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 976.129279] env[62753]: value = "task-1332540" [ 976.129279] env[62753]: _type = "Task" [ 976.129279] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.133415] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332538, 'name': Rename_Task, 'duration_secs': 0.664327} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 976.139585] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 976.139585] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-39268b7d-528b-4643-b7d1-7d07a5fc7439 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.149450] env[62753]: DEBUG oslo_vmware.api [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332540, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.151814] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 976.151814] env[62753]: value = "task-1332541" [ 976.151814] env[62753]: _type = "Task" [ 976.151814] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.161901] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332541, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.346220] env[62753]: DEBUG nova.compute.manager [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 976.471475] env[62753]: DEBUG nova.virt.hardware [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 976.472603] env[62753]: DEBUG nova.virt.hardware [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 976.472776] env[62753]: DEBUG nova.virt.hardware [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 976.472974] env[62753]: DEBUG nova.virt.hardware [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 976.473150] env[62753]: DEBUG nova.virt.hardware [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 976.473308] env[62753]: DEBUG nova.virt.hardware [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 976.473530] env[62753]: DEBUG nova.virt.hardware [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 976.473734] env[62753]: DEBUG nova.virt.hardware [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 976.473956] env[62753]: DEBUG nova.virt.hardware [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 976.474190] env[62753]: DEBUG nova.virt.hardware [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 976.474662] env[62753]: DEBUG nova.virt.hardware [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 976.477356] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3d28ea41-7a84-4a8b-bdea-3f22bc486209 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.487573] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-34645c2d-ea6d-4672-a1b0-234d75e4806c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.495389] env[62753]: DEBUG oslo_vmware.rw_handles [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52669950-1499-f8bd-49c0-ae70a3b18c65/disk-0.vmdk. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 976.496188] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fae2b506-8858-4061-a623-b21f059784f3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.504562] env[62753]: DEBUG oslo_vmware.rw_handles [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52669950-1499-f8bd-49c0-ae70a3b18c65/disk-0.vmdk is in state: ready. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 976.505280] env[62753]: ERROR oslo_vmware.rw_handles [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52669950-1499-f8bd-49c0-ae70a3b18c65/disk-0.vmdk due to incomplete transfer. [ 976.514887] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-0264cebe-08c5-4762-aeb9-c74f8d511ace {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.518819] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.936s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 976.521968] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.147s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 976.527022] env[62753]: INFO nova.compute.claims [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 976.534597] env[62753]: DEBUG oslo_vmware.rw_handles [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52669950-1499-f8bd-49c0-ae70a3b18c65/disk-0.vmdk. {{(pid=62753) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 976.535016] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Uploaded image 3b2f7019-a7ab-4e0e-9ea9-07c8a7334614 to the Glance image server {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 976.537728] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Destroying the VM {{(pid=62753) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 976.538560] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c0857845-6911-4fe1-aa7e-4e58eac539ec {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 976.549970] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 976.549970] env[62753]: value = "task-1332542" [ 976.549970] env[62753]: _type = "Task" [ 976.549970] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 976.563762] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332542, 'name': Destroy_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.645626] env[62753]: DEBUG oslo_vmware.api [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332540, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.647103] env[62753]: INFO nova.scheduler.client.report [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Deleted allocations for instance ff4b295e-2eb2-42dd-b353-08df755f04ea [ 976.672850] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332541, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 976.968276] env[62753]: DEBUG nova.network.neutron [req-483bd152-6b17-4a3b-9ebd-a2dfa9ec67d7 req-ddd63173-0d95-4962-9aa0-7a9f9d748d0a service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Updated VIF entry in instance network info cache for port 12ca481a-0cb3-465e-b04c-802ae3ce6438. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 976.968682] env[62753]: DEBUG nova.network.neutron [req-483bd152-6b17-4a3b-9ebd-a2dfa9ec67d7 req-ddd63173-0d95-4962-9aa0-7a9f9d748d0a service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Updating instance_info_cache with network_info: [{"id": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "address": "fa:16:3e:57:5d:75", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12ca481a-0c", "ovs_interfaceid": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 977.067059] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332542, 'name': Destroy_Task} progress is 33%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.148769] env[62753]: DEBUG oslo_vmware.api [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332540, 'name': ReconfigVM_Task, 'duration_secs': 0.972814} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.149110] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284636', 'volume_id': '368def30-47d9-476e-8669-d2beafd3fba1', 'name': 'volume-368def30-47d9-476e-8669-d2beafd3fba1', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '77cdd901-cc96-4a6f-a696-a65f54a96b1a', 'attached_at': '', 'detached_at': '', 'volume_id': '368def30-47d9-476e-8669-d2beafd3fba1', 'serial': '368def30-47d9-476e-8669-d2beafd3fba1'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 977.164205] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d0aab30-7f21-4a3c-bfdc-4b7299ccfcb8 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "ff4b295e-2eb2-42dd-b353-08df755f04ea" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.844s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.171743] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332541, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.278431] env[62753]: DEBUG nova.network.neutron [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Successfully updated port: c86749ca-c0b7-4c20-a32a-fed94bb47e4d {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 977.332175] env[62753]: DEBUG nova.compute.manager [req-c445414f-5039-4503-856f-76e2e23e89cb req-d465f375-f9f6-463e-a235-89aad7db2ed3 service nova] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Received event network-vif-plugged-c86749ca-c0b7-4c20-a32a-fed94bb47e4d {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 977.333670] env[62753]: DEBUG oslo_concurrency.lockutils [req-c445414f-5039-4503-856f-76e2e23e89cb req-d465f375-f9f6-463e-a235-89aad7db2ed3 service nova] Acquiring lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.333896] env[62753]: DEBUG oslo_concurrency.lockutils [req-c445414f-5039-4503-856f-76e2e23e89cb req-d465f375-f9f6-463e-a235-89aad7db2ed3 service nova] Lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.334126] env[62753]: DEBUG oslo_concurrency.lockutils [req-c445414f-5039-4503-856f-76e2e23e89cb req-d465f375-f9f6-463e-a235-89aad7db2ed3 service nova] Lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 977.334360] env[62753]: DEBUG nova.compute.manager [req-c445414f-5039-4503-856f-76e2e23e89cb req-d465f375-f9f6-463e-a235-89aad7db2ed3 service nova] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] No waiting events found dispatching network-vif-plugged-c86749ca-c0b7-4c20-a32a-fed94bb47e4d {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 977.335272] env[62753]: WARNING nova.compute.manager [req-c445414f-5039-4503-856f-76e2e23e89cb req-d465f375-f9f6-463e-a235-89aad7db2ed3 service nova] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Received unexpected event network-vif-plugged-c86749ca-c0b7-4c20-a32a-fed94bb47e4d for instance with vm_state building and task_state spawning. [ 977.472359] env[62753]: DEBUG oslo_concurrency.lockutils [req-483bd152-6b17-4a3b-9ebd-a2dfa9ec67d7 req-ddd63173-0d95-4962-9aa0-7a9f9d748d0a service nova] Releasing lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 977.564393] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332542, 'name': Destroy_Task, 'duration_secs': 0.579337} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.564688] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Destroyed the VM [ 977.564964] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Deleting Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 977.565262] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-b040dfa9-ba08-423d-a025-16316e7776cd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.577719] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 977.577719] env[62753]: value = "task-1332543" [ 977.577719] env[62753]: _type = "Task" [ 977.577719] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 977.590495] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332543, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 977.667244] env[62753]: DEBUG oslo_vmware.api [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332541, 'name': PowerOnVM_Task, 'duration_secs': 1.387731} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 977.668248] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 977.668248] env[62753]: INFO nova.compute.manager [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Took 10.05 seconds to spawn the instance on the hypervisor. [ 977.668650] env[62753]: DEBUG nova.compute.manager [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 977.672209] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378bcfde-85ef-4dce-a539-4e6d52535b62 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.721139] env[62753]: DEBUG nova.objects.instance [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lazy-loading 'flavor' on Instance uuid 77cdd901-cc96-4a6f-a696-a65f54a96b1a {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 977.762757] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a2b6d8c-1129-46e2-8a7c-42239620d0c7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.772290] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0df50b38-c79f-439e-97c7-22ae473cc008 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.809409] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "refresh_cache-2b90824a-5d49-4ec9-bcf6-4f868ca876cf" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.809409] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "refresh_cache-2b90824a-5d49-4ec9-bcf6-4f868ca876cf" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.809409] env[62753]: DEBUG nova.network.neutron [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 977.812471] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad9b731a-d114-4202-af07-941bfa6ad170 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.822888] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "04fd0022-f20a-4217-9e47-5381635b17a6" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 977.823214] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "04fd0022-f20a-4217-9e47-5381635b17a6" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 977.831997] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6710e6d0-4eca-4a03-8838-f97930cd366e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 977.850447] env[62753]: DEBUG nova.compute.provider_tree [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 977.864902] env[62753]: DEBUG nova.network.neutron [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 977.981683] env[62753]: DEBUG nova.compute.manager [req-b6477a87-e61d-482d-a077-fcbb1e438297 req-1f381f85-e7c8-407a-a11d-0c0e8d18b202 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Received event network-changed-0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 977.982571] env[62753]: DEBUG nova.compute.manager [req-b6477a87-e61d-482d-a077-fcbb1e438297 req-1f381f85-e7c8-407a-a11d-0c0e8d18b202 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Refreshing instance network info cache due to event network-changed-0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 977.982571] env[62753]: DEBUG oslo_concurrency.lockutils [req-b6477a87-e61d-482d-a077-fcbb1e438297 req-1f381f85-e7c8-407a-a11d-0c0e8d18b202 service nova] Acquiring lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 977.982571] env[62753]: DEBUG oslo_concurrency.lockutils [req-b6477a87-e61d-482d-a077-fcbb1e438297 req-1f381f85-e7c8-407a-a11d-0c0e8d18b202 service nova] Acquired lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 977.982571] env[62753]: DEBUG nova.network.neutron [req-b6477a87-e61d-482d-a077-fcbb1e438297 req-1f381f85-e7c8-407a-a11d-0c0e8d18b202 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Refreshing network info cache for port 0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 978.061058] env[62753]: DEBUG nova.network.neutron [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Updating instance_info_cache with network_info: [{"id": "c86749ca-c0b7-4c20-a32a-fed94bb47e4d", "address": "fa:16:3e:8e:d2:8a", "network": {"id": "2823647f-d32d-4a1c-9e02-7577016a260e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1594498922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6c8a6bc946a469fa85cc3ab80d7333a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc86749ca-c0", "ovs_interfaceid": "c86749ca-c0b7-4c20-a32a-fed94bb47e4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.094466] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332543, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.189062] env[62753]: INFO nova.compute.manager [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Took 15.40 seconds to build instance. [ 978.326749] env[62753]: DEBUG nova.compute.manager [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 978.356258] env[62753]: DEBUG nova.scheduler.client.report [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 978.400614] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "b5674964-1928-4ecb-b1fd-8f60a94b4270" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.400833] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "b5674964-1928-4ecb-b1fd-8f60a94b4270" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.401132] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "b5674964-1928-4ecb-b1fd-8f60a94b4270-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.401362] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "b5674964-1928-4ecb-b1fd-8f60a94b4270-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.401526] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "b5674964-1928-4ecb-b1fd-8f60a94b4270-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.405875] env[62753]: INFO nova.compute.manager [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Terminating instance [ 978.408227] env[62753]: DEBUG nova.compute.manager [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 978.408425] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 978.409408] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba580619-ceb0-46de-a250-b37e68ff8a91 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.418097] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 978.420134] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ca618afd-09b9-411d-8227-ad12aa12f4a2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.425388] env[62753]: DEBUG oslo_vmware.api [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 978.425388] env[62753]: value = "task-1332544" [ 978.425388] env[62753]: _type = "Task" [ 978.425388] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.435578] env[62753]: DEBUG oslo_vmware.api [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332544, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.452408] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.565735] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "refresh_cache-2b90824a-5d49-4ec9-bcf6-4f868ca876cf" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 978.566122] env[62753]: DEBUG nova.compute.manager [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Instance network_info: |[{"id": "c86749ca-c0b7-4c20-a32a-fed94bb47e4d", "address": "fa:16:3e:8e:d2:8a", "network": {"id": "2823647f-d32d-4a1c-9e02-7577016a260e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1594498922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6c8a6bc946a469fa85cc3ab80d7333a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc86749ca-c0", "ovs_interfaceid": "c86749ca-c0b7-4c20-a32a-fed94bb47e4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 978.566561] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:d2:8a', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9d39252e-42ef-4252-98d3-62af5a0d109d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'c86749ca-c0b7-4c20-a32a-fed94bb47e4d', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 978.577020] env[62753]: DEBUG oslo.service.loopingcall [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 978.577312] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 978.577545] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-283eb7a7-84fe-4093-a441-83ba7ce1553c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.611249] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332543, 'name': RemoveSnapshot_Task, 'duration_secs': 0.573159} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.611249] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Deleted Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 978.611533] env[62753]: DEBUG nova.compute.manager [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 978.611837] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 978.611837] env[62753]: value = "task-1332545" [ 978.611837] env[62753]: _type = "Task" [ 978.611837] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 978.612573] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f08613f-fdf6-4b23-b230-62b93c7d0d2a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 978.631035] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332545, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 978.690863] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6b7296d-2abd-4b1f-b2b4-68098564b862 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "bba5e3cb-c896-4ce7-ac7f-1cf2aba20465" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.914s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.730584] env[62753]: DEBUG oslo_concurrency.lockutils [None req-793d17ca-46ec-448b-b53c-a88d9d2277ab tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 4.828s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.731823] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.280s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.763417] env[62753]: DEBUG nova.network.neutron [req-b6477a87-e61d-482d-a077-fcbb1e438297 req-1f381f85-e7c8-407a-a11d-0c0e8d18b202 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Updated VIF entry in instance network info cache for port 0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 978.763592] env[62753]: DEBUG nova.network.neutron [req-b6477a87-e61d-482d-a077-fcbb1e438297 req-1f381f85-e7c8-407a-a11d-0c0e8d18b202 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Updating instance_info_cache with network_info: [{"id": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "address": "fa:16:3e:6b:34:1e", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cd05bfa-82", "ovs_interfaceid": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 978.848256] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 978.861495] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.340s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 978.862347] env[62753]: DEBUG nova.compute.manager [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 978.865320] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.180s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 978.865583] env[62753]: DEBUG nova.objects.instance [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lazy-loading 'resources' on Instance uuid 8066aeb2-7be4-46e3-aba6-124fef342a49 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 978.936673] env[62753]: DEBUG oslo_vmware.api [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332544, 'name': PowerOffVM_Task, 'duration_secs': 0.290345} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 978.937048] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 978.937161] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 978.937397] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-869c9507-f88f-410c-9991-ea07d2a68064 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.045879] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 979.046268] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 979.046499] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Deleting the datastore file [datastore1] b5674964-1928-4ecb-b1fd-8f60a94b4270 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 979.046933] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-41e6dede-3fc9-414e-9883-29eb64a0c20c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.058227] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "e1c59139-ff02-453e-b54f-a56b1e40fdb5" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 979.058534] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "e1c59139-ff02-453e-b54f-a56b1e40fdb5" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 979.059857] env[62753]: DEBUG oslo_vmware.api [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 979.059857] env[62753]: value = "task-1332547" [ 979.059857] env[62753]: _type = "Task" [ 979.059857] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.070253] env[62753]: DEBUG oslo_vmware.api [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332547, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.126609] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332545, 'name': CreateVM_Task, 'duration_secs': 0.427252} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.126794] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 979.127538] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.127713] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.128106] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 979.128858] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a7e26cfe-1dcf-4c4e-ad1f-eb21dd901278 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.135767] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 979.135767] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52086653-4fc8-5ba6-0370-006ac6444438" [ 979.135767] env[62753]: _type = "Task" [ 979.135767] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.136310] env[62753]: INFO nova.compute.manager [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Shelve offloading [ 979.138387] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 979.141996] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-696f59d6-c315-40ab-be90-a418bbbcf590 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.151789] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52086653-4fc8-5ba6-0370-006ac6444438, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.153585] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 979.153585] env[62753]: value = "task-1332548" [ 979.153585] env[62753]: _type = "Task" [ 979.153585] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.164825] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] VM already powered off {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 979.165052] env[62753]: DEBUG nova.compute.manager [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 979.165883] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c231f5f7-607d-4ba2-8664-c7204efa87dd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.170079] env[62753]: INFO nova.compute.manager [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Rescuing [ 979.170351] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "refresh_cache-bba5e3cb-c896-4ce7-ac7f-1cf2aba20465" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.170533] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquired lock "refresh_cache-bba5e3cb-c896-4ce7-ac7f-1cf2aba20465" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.170742] env[62753]: DEBUG nova.network.neutron [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 979.173582] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.173776] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquired lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.173963] env[62753]: DEBUG nova.network.neutron [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 979.235429] env[62753]: INFO nova.compute.manager [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Detaching volume 36f91b11-b0aa-4304-aa25-a7ff2fa37218 [ 979.266885] env[62753]: DEBUG oslo_concurrency.lockutils [req-b6477a87-e61d-482d-a077-fcbb1e438297 req-1f381f85-e7c8-407a-a11d-0c0e8d18b202 service nova] Releasing lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.269607] env[62753]: INFO nova.virt.block_device [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Attempting to driver detach volume 36f91b11-b0aa-4304-aa25-a7ff2fa37218 from mountpoint /dev/sdc [ 979.269840] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Volume detach. Driver type: vmdk {{(pid=62753) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 979.270046] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284643', 'volume_id': '36f91b11-b0aa-4304-aa25-a7ff2fa37218', 'name': 'volume-36f91b11-b0aa-4304-aa25-a7ff2fa37218', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '77cdd901-cc96-4a6f-a696-a65f54a96b1a', 'attached_at': '', 'detached_at': '', 'volume_id': '36f91b11-b0aa-4304-aa25-a7ff2fa37218', 'serial': '36f91b11-b0aa-4304-aa25-a7ff2fa37218'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 979.270964] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-705e867b-fc1a-4320-b213-b27f6e0b3e70 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.293788] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c05c94b-53f6-4175-8152-4190f1d98a9e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.302389] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b41ec960-19d0-49bd-9218-23d428186c48 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.327303] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e2481e-46b5-45f4-a8c0-e84d598f0ae7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.345404] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] The volume has not been displaced from its original location: [datastore1] volume-36f91b11-b0aa-4304-aa25-a7ff2fa37218/volume-36f91b11-b0aa-4304-aa25-a7ff2fa37218.vmdk. No consolidation needed. {{(pid=62753) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 979.351168] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Reconfiguring VM instance instance-00000049 to detach disk 2002 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 979.351598] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b855189c-57aa-4486-bedc-05b37778259c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.368083] env[62753]: DEBUG nova.compute.manager [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Received event network-changed-c86749ca-c0b7-4c20-a32a-fed94bb47e4d {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 979.369066] env[62753]: DEBUG nova.compute.manager [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Refreshing instance network info cache due to event network-changed-c86749ca-c0b7-4c20-a32a-fed94bb47e4d. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 979.369066] env[62753]: DEBUG oslo_concurrency.lockutils [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] Acquiring lock "refresh_cache-2b90824a-5d49-4ec9-bcf6-4f868ca876cf" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.369066] env[62753]: DEBUG oslo_concurrency.lockutils [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] Acquired lock "refresh_cache-2b90824a-5d49-4ec9-bcf6-4f868ca876cf" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.369066] env[62753]: DEBUG nova.network.neutron [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Refreshing network info cache for port c86749ca-c0b7-4c20-a32a-fed94bb47e4d {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 979.371201] env[62753]: DEBUG nova.compute.utils [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 979.376377] env[62753]: DEBUG nova.compute.manager [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 979.376377] env[62753]: DEBUG nova.network.neutron [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 979.379409] env[62753]: DEBUG oslo_vmware.api [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 979.379409] env[62753]: value = "task-1332549" [ 979.379409] env[62753]: _type = "Task" [ 979.379409] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.390288] env[62753]: DEBUG oslo_vmware.api [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332549, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.428173] env[62753]: DEBUG nova.policy [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae52823f7c93454e8089261c3ca44321', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d52d6eaee934be5ab0e0003df1ce316', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 979.562402] env[62753]: DEBUG nova.compute.manager [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 979.575903] env[62753]: DEBUG oslo_vmware.api [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332547, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.259158} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.576322] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 979.576517] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 979.576766] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 979.576991] env[62753]: INFO nova.compute.manager [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Took 1.17 seconds to destroy the instance on the hypervisor. [ 979.577313] env[62753]: DEBUG oslo.service.loopingcall [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 979.577566] env[62753]: DEBUG nova.compute.manager [-] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 979.577689] env[62753]: DEBUG nova.network.neutron [-] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 979.581142] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082e95fb-5ef8-42f9-8368-66ffdb4ee263 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.590069] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adf6d642-ce9e-486d-aea4-bb06f387b518 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.625140] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-86f401ee-fae0-4d8f-9193-c98e36bdb27a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.634606] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0cab969-b4d6-47ef-b894-66239777f3f2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.657545] env[62753]: DEBUG nova.compute.provider_tree [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 979.663033] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52086653-4fc8-5ba6-0370-006ac6444438, 'name': SearchDatastore_Task, 'duration_secs': 0.019787} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.663597] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 979.665369] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 979.665369] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 979.665369] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 979.665369] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 979.665369] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff3ea8ff-050b-407a-8da4-2e4121f81ce7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.680245] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 979.680474] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 979.683643] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1b670410-fa0e-41d6-ae00-977336d838c9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.691840] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 979.691840] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52748692-92d2-c1f7-200e-b2d5c83f5260" [ 979.691840] env[62753]: _type = "Task" [ 979.691840] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.702095] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52748692-92d2-c1f7-200e-b2d5c83f5260, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 979.735968] env[62753]: DEBUG nova.network.neutron [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Successfully created port: 59ca12e1-f9c6-4e17-a496-248b15733019 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 979.876632] env[62753]: DEBUG nova.compute.manager [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 979.901278] env[62753]: DEBUG oslo_vmware.api [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332549, 'name': ReconfigVM_Task, 'duration_secs': 0.396599} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 979.901693] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Reconfigured VM instance instance-00000049 to detach disk 2002 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 979.907467] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a6c8f4b6-3e6c-43e5-9e85-0792b63ebe93 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 979.926790] env[62753]: DEBUG oslo_vmware.api [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 979.926790] env[62753]: value = "task-1332550" [ 979.926790] env[62753]: _type = "Task" [ 979.926790] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 979.940802] env[62753]: DEBUG oslo_vmware.api [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332550, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.035226] env[62753]: DEBUG nova.compute.manager [req-b6bef371-d37b-4955-abb5-07e343d269b3 req-5fa72002-f4e4-4dbc-a391-0722e06049da service nova] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Received event network-vif-deleted-d17809a7-cdfb-44c8-9b83-1d5110ede0d3 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 980.035558] env[62753]: INFO nova.compute.manager [req-b6bef371-d37b-4955-abb5-07e343d269b3 req-5fa72002-f4e4-4dbc-a391-0722e06049da service nova] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Neutron deleted interface d17809a7-cdfb-44c8-9b83-1d5110ede0d3; detaching it from the instance and deleting it from the info cache [ 980.035880] env[62753]: DEBUG nova.network.neutron [req-b6bef371-d37b-4955-abb5-07e343d269b3 req-5fa72002-f4e4-4dbc-a391-0722e06049da service nova] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.077336] env[62753]: DEBUG nova.network.neutron [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Updating instance_info_cache with network_info: [{"id": "4832b8bd-7a18-4f8e-a05b-7421417d64ff", "address": "fa:16:3e:a5:4c:97", "network": {"id": "22a90636-182f-4944-893f-329f446b07c4", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1513076888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "37368e15fa5a4a0a97ac1c22fcf0bb80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap4832b8bd-7a", "ovs_interfaceid": "4832b8bd-7a18-4f8e-a05b-7421417d64ff", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.093460] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 980.165105] env[62753]: DEBUG nova.scheduler.client.report [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 980.203503] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52748692-92d2-c1f7-200e-b2d5c83f5260, 'name': SearchDatastore_Task, 'duration_secs': 0.013011} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.204459] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-016b437b-c6a1-4673-b18e-6fa7ac44a9fc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.211212] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 980.211212] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52271339-39d3-6468-d5f9-2b2f83ba85c0" [ 980.211212] env[62753]: _type = "Task" [ 980.211212] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.221411] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52271339-39d3-6468-d5f9-2b2f83ba85c0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.350135] env[62753]: DEBUG nova.network.neutron [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Updating instance_info_cache with network_info: [{"id": "9c52b1e4-82d2-4449-91f3-a624596bdc79", "address": "fa:16:3e:8e:1b:de", "network": {"id": "12446a5e-7329-4f8f-aa28-a8b9d78e7573", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1766107072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "371a5e8b16be4cffbdeb2139757684fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7c7f16b-a2e3-4d1f-9b7e-a44a9ebb589c", "external-id": "nsx-vlan-transportzone-772", "segmentation_id": 772, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c52b1e4-82", "ovs_interfaceid": "9c52b1e4-82d2-4449-91f3-a624596bdc79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.445912] env[62753]: DEBUG oslo_vmware.api [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332550, 'name': ReconfigVM_Task, 'duration_secs': 0.180093} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.447053] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284643', 'volume_id': '36f91b11-b0aa-4304-aa25-a7ff2fa37218', 'name': 'volume-36f91b11-b0aa-4304-aa25-a7ff2fa37218', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '77cdd901-cc96-4a6f-a696-a65f54a96b1a', 'attached_at': '', 'detached_at': '', 'volume_id': '36f91b11-b0aa-4304-aa25-a7ff2fa37218', 'serial': '36f91b11-b0aa-4304-aa25-a7ff2fa37218'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 980.507774] env[62753]: DEBUG nova.network.neutron [-] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.527424] env[62753]: DEBUG nova.network.neutron [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Updated VIF entry in instance network info cache for port c86749ca-c0b7-4c20-a32a-fed94bb47e4d. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 980.528109] env[62753]: DEBUG nova.network.neutron [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Updating instance_info_cache with network_info: [{"id": "c86749ca-c0b7-4c20-a32a-fed94bb47e4d", "address": "fa:16:3e:8e:d2:8a", "network": {"id": "2823647f-d32d-4a1c-9e02-7577016a260e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1594498922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6c8a6bc946a469fa85cc3ab80d7333a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc86749ca-c0", "ovs_interfaceid": "c86749ca-c0b7-4c20-a32a-fed94bb47e4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 980.538935] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-afdf2578-1fb8-4300-a374-794875b8e3d7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.550353] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-887fb0a2-5a3f-462a-98ad-b6319828dee0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.583731] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Releasing lock "refresh_cache-bba5e3cb-c896-4ce7-ac7f-1cf2aba20465" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.585884] env[62753]: DEBUG nova.compute.manager [req-b6bef371-d37b-4955-abb5-07e343d269b3 req-5fa72002-f4e4-4dbc-a391-0722e06049da service nova] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Detach interface failed, port_id=d17809a7-cdfb-44c8-9b83-1d5110ede0d3, reason: Instance b5674964-1928-4ecb-b1fd-8f60a94b4270 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 980.671104] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.806s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 980.673505] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.825s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 980.675453] env[62753]: INFO nova.compute.claims [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 980.692558] env[62753]: INFO nova.scheduler.client.report [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Deleted allocations for instance 8066aeb2-7be4-46e3-aba6-124fef342a49 [ 980.725861] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52271339-39d3-6468-d5f9-2b2f83ba85c0, 'name': SearchDatastore_Task, 'duration_secs': 0.012722} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 980.726170] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.726430] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 2b90824a-5d49-4ec9-bcf6-4f868ca876cf/2b90824a-5d49-4ec9-bcf6-4f868ca876cf.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 980.726699] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2ac15219-dd4d-47cb-ac52-84544c5d32dc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.734244] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 980.734244] env[62753]: value = "task-1332551" [ 980.734244] env[62753]: _type = "Task" [ 980.734244] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 980.744132] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332551, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 980.855675] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Releasing lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 980.895281] env[62753]: DEBUG nova.compute.manager [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 980.919888] env[62753]: DEBUG nova.virt.hardware [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 980.920363] env[62753]: DEBUG nova.virt.hardware [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 980.920670] env[62753]: DEBUG nova.virt.hardware [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 980.921018] env[62753]: DEBUG nova.virt.hardware [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 980.921312] env[62753]: DEBUG nova.virt.hardware [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 980.921602] env[62753]: DEBUG nova.virt.hardware [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 980.921980] env[62753]: DEBUG nova.virt.hardware [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 980.922299] env[62753]: DEBUG nova.virt.hardware [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 980.922622] env[62753]: DEBUG nova.virt.hardware [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 980.922948] env[62753]: DEBUG nova.virt.hardware [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 980.923293] env[62753]: DEBUG nova.virt.hardware [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 980.924672] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3b9b763-720e-46f1-a494-8d847abd15b0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.940038] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7287e52-3ea5-4e53-a98c-212130147d94 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 980.992665] env[62753]: DEBUG nova.objects.instance [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lazy-loading 'flavor' on Instance uuid 77cdd901-cc96-4a6f-a696-a65f54a96b1a {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 981.013541] env[62753]: INFO nova.compute.manager [-] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Took 1.43 seconds to deallocate network for instance. [ 981.031373] env[62753]: DEBUG oslo_concurrency.lockutils [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] Releasing lock "refresh_cache-2b90824a-5d49-4ec9-bcf6-4f868ca876cf" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 981.031610] env[62753]: DEBUG nova.compute.manager [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Received event network-changed-0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.031793] env[62753]: DEBUG nova.compute.manager [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Refreshing instance network info cache due to event network-changed-0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 981.032091] env[62753]: DEBUG oslo_concurrency.lockutils [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] Acquiring lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.032193] env[62753]: DEBUG oslo_concurrency.lockutils [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] Acquired lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.032376] env[62753]: DEBUG nova.network.neutron [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Refreshing network info cache for port 0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 981.116379] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 981.116752] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ced50f8e-1258-4c72-bfed-003f3e9033e2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.126673] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 981.126673] env[62753]: value = "task-1332552" [ 981.126673] env[62753]: _type = "Task" [ 981.126673] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.151621] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332552, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.184365] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 981.185367] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d363b0d-8533-488f-a8bd-026f44b3f8c8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.195531] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 981.195858] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-959aa572-bff1-47a1-b981-abe225275a88 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.201876] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea9ab782-c43a-4a4c-bdcf-9d43d3f54613 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "8066aeb2-7be4-46e3-aba6-124fef342a49" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.553s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.245731] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332551, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.316061] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 981.316448] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 981.316710] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Deleting the datastore file [datastore2] 4d20e083-2959-453a-8875-47955bc02613 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 981.317027] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-563a1f3c-964f-4336-8a9c-331ec4c4dc42 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.326793] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 981.326793] env[62753]: value = "task-1332554" [ 981.326793] env[62753]: _type = "Task" [ 981.326793] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.337271] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332554, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.397416] env[62753]: DEBUG nova.compute.manager [req-d2bfd8f1-2bf8-447d-b6ea-b1fb846f6fa7 req-df018354-cf36-45d7-bf15-5ae556165748 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Received event network-vif-unplugged-9c52b1e4-82d2-4449-91f3-a624596bdc79 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.397786] env[62753]: DEBUG oslo_concurrency.lockutils [req-d2bfd8f1-2bf8-447d-b6ea-b1fb846f6fa7 req-df018354-cf36-45d7-bf15-5ae556165748 service nova] Acquiring lock "4d20e083-2959-453a-8875-47955bc02613-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.398148] env[62753]: DEBUG oslo_concurrency.lockutils [req-d2bfd8f1-2bf8-447d-b6ea-b1fb846f6fa7 req-df018354-cf36-45d7-bf15-5ae556165748 service nova] Lock "4d20e083-2959-453a-8875-47955bc02613-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.398470] env[62753]: DEBUG oslo_concurrency.lockutils [req-d2bfd8f1-2bf8-447d-b6ea-b1fb846f6fa7 req-df018354-cf36-45d7-bf15-5ae556165748 service nova] Lock "4d20e083-2959-453a-8875-47955bc02613-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.398788] env[62753]: DEBUG nova.compute.manager [req-d2bfd8f1-2bf8-447d-b6ea-b1fb846f6fa7 req-df018354-cf36-45d7-bf15-5ae556165748 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] No waiting events found dispatching network-vif-unplugged-9c52b1e4-82d2-4449-91f3-a624596bdc79 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 981.399140] env[62753]: WARNING nova.compute.manager [req-d2bfd8f1-2bf8-447d-b6ea-b1fb846f6fa7 req-df018354-cf36-45d7-bf15-5ae556165748 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Received unexpected event network-vif-unplugged-9c52b1e4-82d2-4449-91f3-a624596bdc79 for instance with vm_state shelved and task_state shelving_offloading. [ 981.399879] env[62753]: DEBUG nova.compute.manager [req-d2bfd8f1-2bf8-447d-b6ea-b1fb846f6fa7 req-df018354-cf36-45d7-bf15-5ae556165748 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Received event network-changed-9c52b1e4-82d2-4449-91f3-a624596bdc79 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.399879] env[62753]: DEBUG nova.compute.manager [req-d2bfd8f1-2bf8-447d-b6ea-b1fb846f6fa7 req-df018354-cf36-45d7-bf15-5ae556165748 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Refreshing instance network info cache due to event network-changed-9c52b1e4-82d2-4449-91f3-a624596bdc79. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 981.400093] env[62753]: DEBUG oslo_concurrency.lockutils [req-d2bfd8f1-2bf8-447d-b6ea-b1fb846f6fa7 req-df018354-cf36-45d7-bf15-5ae556165748 service nova] Acquiring lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.400351] env[62753]: DEBUG oslo_concurrency.lockutils [req-d2bfd8f1-2bf8-447d-b6ea-b1fb846f6fa7 req-df018354-cf36-45d7-bf15-5ae556165748 service nova] Acquired lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.401033] env[62753]: DEBUG nova.network.neutron [req-d2bfd8f1-2bf8-447d-b6ea-b1fb846f6fa7 req-df018354-cf36-45d7-bf15-5ae556165748 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Refreshing network info cache for port 9c52b1e4-82d2-4449-91f3-a624596bdc79 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 981.519579] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.639999] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332552, 'name': PowerOffVM_Task, 'duration_secs': 0.244613} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.640364] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 981.641128] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bbcacd4e-ab2a-4d43-a94e-44abcfd5ab4d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.664977] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48811a20-2fd1-4f74-939d-091f3a763f7b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.721107] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 981.721622] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f658ce8b-4445-43ce-bf5d-2e67ef1a8ced {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.730675] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 981.730675] env[62753]: value = "task-1332555" [ 981.730675] env[62753]: _type = "Task" [ 981.730675] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.754231] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] VM already powered off {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 981.754231] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 981.754498] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 981.754747] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 981.755016] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 981.759512] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-48f740a6-a7a1-4017-b15e-2bd5bafa56e9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.763321] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332551, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.759676} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.764219] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 2b90824a-5d49-4ec9-bcf6-4f868ca876cf/2b90824a-5d49-4ec9-bcf6-4f868ca876cf.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 981.764219] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 981.764565] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-7c5cbada-3254-454d-a029-f0f506ce585b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.772548] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 981.772803] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 981.773623] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dedf6cd3-cba5-4a08-81fc-e07de49863b7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.777692] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 981.777692] env[62753]: value = "task-1332556" [ 981.777692] env[62753]: _type = "Task" [ 981.777692] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.783983] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 981.783983] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]526103fa-8d63-5707-ad37-069d95031b22" [ 981.783983] env[62753]: _type = "Task" [ 981.783983] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.796041] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332556, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.807381] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]526103fa-8d63-5707-ad37-069d95031b22, 'name': SearchDatastore_Task, 'duration_secs': 0.016678} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.808498] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0ae14baa-508b-4b7e-87ca-90db3b59fb66 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.815837] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 981.815837] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52cc6dfe-7777-ab91-6414-9881e43af95b" [ 981.815837] env[62753]: _type = "Task" [ 981.815837] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 981.829177] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52cc6dfe-7777-ab91-6414-9881e43af95b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 981.840818] env[62753]: DEBUG oslo_vmware.api [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332554, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.412267} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 981.841230] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 981.841510] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 981.841785] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 981.868847] env[62753]: INFO nova.scheduler.client.report [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Deleted allocations for instance 4d20e083-2959-453a-8875-47955bc02613 [ 981.905863] env[62753]: DEBUG nova.network.neutron [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Updated VIF entry in instance network info cache for port 0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 981.905863] env[62753]: DEBUG nova.network.neutron [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Updating instance_info_cache with network_info: [{"id": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "address": "fa:16:3e:6b:34:1e", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cd05bfa-82", "ovs_interfaceid": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 981.927742] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c27a3e-47e3-4a41-abd0-a7e618cd26a6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.938193] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d964a7f0-625d-4281-92c9-bef183756d19 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.943889] env[62753]: DEBUG nova.compute.manager [req-0722bf4d-3899-4799-8745-639cd9939318 req-3d4b52b4-d816-48ec-823a-c846b1d97dac service nova] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Received event network-vif-plugged-59ca12e1-f9c6-4e17-a496-248b15733019 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 981.944131] env[62753]: DEBUG oslo_concurrency.lockutils [req-0722bf4d-3899-4799-8745-639cd9939318 req-3d4b52b4-d816-48ec-823a-c846b1d97dac service nova] Acquiring lock "91fc519e-c9bb-4605-bdc6-4f463561b988-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 981.944373] env[62753]: DEBUG oslo_concurrency.lockutils [req-0722bf4d-3899-4799-8745-639cd9939318 req-3d4b52b4-d816-48ec-823a-c846b1d97dac service nova] Lock "91fc519e-c9bb-4605-bdc6-4f463561b988-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 981.944553] env[62753]: DEBUG oslo_concurrency.lockutils [req-0722bf4d-3899-4799-8745-639cd9939318 req-3d4b52b4-d816-48ec-823a-c846b1d97dac service nova] Lock "91fc519e-c9bb-4605-bdc6-4f463561b988-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 981.944728] env[62753]: DEBUG nova.compute.manager [req-0722bf4d-3899-4799-8745-639cd9939318 req-3d4b52b4-d816-48ec-823a-c846b1d97dac service nova] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] No waiting events found dispatching network-vif-plugged-59ca12e1-f9c6-4e17-a496-248b15733019 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 981.944941] env[62753]: WARNING nova.compute.manager [req-0722bf4d-3899-4799-8745-639cd9939318 req-3d4b52b4-d816-48ec-823a-c846b1d97dac service nova] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Received unexpected event network-vif-plugged-59ca12e1-f9c6-4e17-a496-248b15733019 for instance with vm_state building and task_state spawning. [ 981.976093] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b92f135-175e-4abb-a73a-e3a050270c67 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 981.985910] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78b07c7c-fdb0-402b-aca7-7231a12d5699 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.002208] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c85d6754-7540-411a-9a8f-43e9738d2010 tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.270s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.003462] env[62753]: DEBUG nova.compute.provider_tree [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 982.164937] env[62753]: DEBUG nova.network.neutron [req-d2bfd8f1-2bf8-447d-b6ea-b1fb846f6fa7 req-df018354-cf36-45d7-bf15-5ae556165748 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Updated VIF entry in instance network info cache for port 9c52b1e4-82d2-4449-91f3-a624596bdc79. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 982.164937] env[62753]: DEBUG nova.network.neutron [req-d2bfd8f1-2bf8-447d-b6ea-b1fb846f6fa7 req-df018354-cf36-45d7-bf15-5ae556165748 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Updating instance_info_cache with network_info: [{"id": "9c52b1e4-82d2-4449-91f3-a624596bdc79", "address": "fa:16:3e:8e:1b:de", "network": {"id": "12446a5e-7329-4f8f-aa28-a8b9d78e7573", "bridge": null, "label": "tempest-ServersNegativeTestJSON-1766107072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "371a5e8b16be4cffbdeb2139757684fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tap9c52b1e4-82", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 982.289423] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332556, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083607} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.289938] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 982.290872] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6467823e-cc20-46d2-bcb1-bc0d440f28a8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.328912] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 2b90824a-5d49-4ec9-bcf6-4f868ca876cf/2b90824a-5d49-4ec9-bcf6-4f868ca876cf.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 982.329506] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4b4e8e1-5744-48c0-a0d9-89d5e38a8432 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.369590] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52cc6dfe-7777-ab91-6414-9881e43af95b, 'name': SearchDatastore_Task, 'duration_secs': 0.01467} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.371593] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.371991] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Copying virtual disk from [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] bba5e3cb-c896-4ce7-ac7f-1cf2aba20465/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk. {{(pid=62753) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 982.372502] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 982.372502] env[62753]: value = "task-1332557" [ 982.372502] env[62753]: _type = "Task" [ 982.372502] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.373531] env[62753]: DEBUG nova.network.neutron [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Successfully updated port: 59ca12e1-f9c6-4e17-a496-248b15733019 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 982.376043] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.376491] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-13d8c8e6-6af3-43cd-864b-ddb03f8e10e7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.394444] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332557, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.396604] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 982.396604] env[62753]: value = "task-1332558" [ 982.396604] env[62753]: _type = "Task" [ 982.396604] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.408771] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332558, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.410444] env[62753]: DEBUG oslo_concurrency.lockutils [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] Releasing lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.410735] env[62753]: DEBUG nova.compute.manager [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Received event network-changed-12ca481a-0cb3-465e-b04c-802ae3ce6438 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 982.410915] env[62753]: DEBUG nova.compute.manager [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Refreshing instance network info cache due to event network-changed-12ca481a-0cb3-465e-b04c-802ae3ce6438. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 982.411598] env[62753]: DEBUG oslo_concurrency.lockutils [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] Acquiring lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.411598] env[62753]: DEBUG oslo_concurrency.lockutils [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] Acquired lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.411598] env[62753]: DEBUG nova.network.neutron [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Refreshing network info cache for port 12ca481a-0cb3-465e-b04c-802ae3ce6438 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 982.506992] env[62753]: DEBUG nova.scheduler.client.report [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 982.667667] env[62753]: DEBUG oslo_concurrency.lockutils [req-d2bfd8f1-2bf8-447d-b6ea-b1fb846f6fa7 req-df018354-cf36-45d7-bf15-5ae556165748 service nova] Releasing lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 982.696021] env[62753]: DEBUG oslo_concurrency.lockutils [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "3009b1e4-53b1-4b12-8aaf-3c36b37c5d99" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.696021] env[62753]: DEBUG oslo_concurrency.lockutils [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "3009b1e4-53b1-4b12-8aaf-3c36b37c5d99" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.880503] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "refresh_cache-91fc519e-c9bb-4605-bdc6-4f463561b988" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 982.880668] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "refresh_cache-91fc519e-c9bb-4605-bdc6-4f463561b988" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 982.880825] env[62753]: DEBUG nova.network.neutron [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 982.893277] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332557, 'name': ReconfigVM_Task, 'duration_secs': 0.373574} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 982.894324] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 2b90824a-5d49-4ec9-bcf6-4f868ca876cf/2b90824a-5d49-4ec9-bcf6-4f868ca876cf.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 982.894999] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-87ead1e7-c56e-4d75-bfc7-d2c68d679709 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.905210] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 982.905210] env[62753]: value = "task-1332559" [ 982.905210] env[62753]: _type = "Task" [ 982.905210] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.912173] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332558, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.920361] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.920651] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.920861] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 982.921060] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 982.921238] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 982.923102] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332559, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 982.924147] env[62753]: INFO nova.compute.manager [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Terminating instance [ 982.926403] env[62753]: DEBUG nova.compute.manager [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 982.926640] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 982.927629] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3885476-cf19-47a2-94a3-ded9b5b92b28 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.938662] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 982.938988] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9c128dbb-4968-484c-9caa-a90e3d8ed015 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 982.950793] env[62753]: DEBUG oslo_vmware.api [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 982.950793] env[62753]: value = "task-1332560" [ 982.950793] env[62753]: _type = "Task" [ 982.950793] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 982.967856] env[62753]: DEBUG oslo_vmware.api [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332560, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.012892] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.339s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 983.013758] env[62753]: DEBUG nova.compute.manager [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 983.018418] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 2.925s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 983.020246] env[62753]: INFO nova.compute.claims [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 983.198442] env[62753]: DEBUG nova.compute.manager [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 983.246805] env[62753]: DEBUG nova.network.neutron [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Updated VIF entry in instance network info cache for port 12ca481a-0cb3-465e-b04c-802ae3ce6438. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 983.247219] env[62753]: DEBUG nova.network.neutron [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Updating instance_info_cache with network_info: [{"id": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "address": "fa:16:3e:57:5d:75", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12ca481a-0c", "ovs_interfaceid": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.409293] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332558, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.671837} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.411573] env[62753]: INFO nova.virt.vmwareapi.ds_util [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Copied virtual disk from [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] bba5e3cb-c896-4ce7-ac7f-1cf2aba20465/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk. [ 983.412422] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e403fe8-e813-4d1c-8dd6-9d30071f6909 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.421953] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332559, 'name': Rename_Task, 'duration_secs': 0.249052} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.438582] env[62753]: DEBUG nova.network.neutron [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 983.440834] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 983.449550] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Reconfiguring VM instance instance-00000055 to attach disk [datastore1] bba5e3cb-c896-4ce7-ac7f-1cf2aba20465/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 983.451239] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-203bab1c-2233-48df-9696-2c05907f4508 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.453290] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a5a59ad6-cb19-4c71-8ae8-5d5fb1fd40f8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.470529] env[62753]: DEBUG nova.compute.manager [req-6c115c1d-7423-4fc0-a50f-df0779e3de20 req-9cfd48ef-5f3c-4dab-9fff-5accbf5bca78 service nova] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Received event network-changed-59ca12e1-f9c6-4e17-a496-248b15733019 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 983.470728] env[62753]: DEBUG nova.compute.manager [req-6c115c1d-7423-4fc0-a50f-df0779e3de20 req-9cfd48ef-5f3c-4dab-9fff-5accbf5bca78 service nova] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Refreshing instance network info cache due to event network-changed-59ca12e1-f9c6-4e17-a496-248b15733019. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 983.470933] env[62753]: DEBUG oslo_concurrency.lockutils [req-6c115c1d-7423-4fc0-a50f-df0779e3de20 req-9cfd48ef-5f3c-4dab-9fff-5accbf5bca78 service nova] Acquiring lock "refresh_cache-91fc519e-c9bb-4605-bdc6-4f463561b988" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 983.482236] env[62753]: DEBUG oslo_vmware.api [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332560, 'name': PowerOffVM_Task, 'duration_secs': 0.328956} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 983.486431] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 983.486624] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 983.486950] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 983.486950] env[62753]: value = "task-1332561" [ 983.486950] env[62753]: _type = "Task" [ 983.486950] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.487208] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 983.487208] env[62753]: value = "task-1332562" [ 983.487208] env[62753]: _type = "Task" [ 983.487208] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.487415] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-c7978d4d-3637-4800-a91e-7f6f2e4158b4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.501018] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332562, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.504423] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332561, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.525607] env[62753]: DEBUG nova.compute.utils [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 983.530274] env[62753]: DEBUG nova.compute.manager [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 983.530486] env[62753]: DEBUG nova.network.neutron [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 983.578465] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 983.578786] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 983.579120] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Deleting the datastore file [datastore2] 77cdd901-cc96-4a6f-a696-a65f54a96b1a {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 983.579496] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-438347be-5054-482a-b161-477b0ce0644c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 983.589588] env[62753]: DEBUG oslo_vmware.api [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for the task: (returnval){ [ 983.589588] env[62753]: value = "task-1332564" [ 983.589588] env[62753]: _type = "Task" [ 983.589588] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 983.599762] env[62753]: DEBUG oslo_vmware.api [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332564, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 983.601373] env[62753]: DEBUG nova.policy [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '592d1d366cc4461299dbc28cee63e5b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8925b84dcf9a47fbaf2eb8044b3850fa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 983.699573] env[62753]: DEBUG nova.network.neutron [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Updating instance_info_cache with network_info: [{"id": "59ca12e1-f9c6-4e17-a496-248b15733019", "address": "fa:16:3e:3c:35:4c", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59ca12e1-f9", "ovs_interfaceid": "59ca12e1-f9c6-4e17-a496-248b15733019", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 983.724424] env[62753]: DEBUG oslo_concurrency.lockutils [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 983.750561] env[62753]: DEBUG oslo_concurrency.lockutils [req-318e4cf3-8f2c-43a6-8296-4c2b703ec8c4 req-ba024143-bafd-4e87-8ef1-cc7a0804d228 service nova] Releasing lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 983.953566] env[62753]: DEBUG nova.network.neutron [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Successfully created port: 011eae16-b648-4e2e-abfb-e424e05e41c2 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 984.003310] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332561, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.006621] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332562, 'name': ReconfigVM_Task, 'duration_secs': 0.413678} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.006897] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Reconfigured VM instance instance-00000055 to attach disk [datastore1] bba5e3cb-c896-4ce7-ac7f-1cf2aba20465/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 984.007796] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5efe8932-ce56-4745-928f-c78c044e1f9b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.030019] env[62753]: DEBUG nova.compute.manager [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 984.042100] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-0e2eff6d-94bd-486c-a5b1-73e3221c0ebe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.061264] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 984.061264] env[62753]: value = "task-1332565" [ 984.061264] env[62753]: _type = "Task" [ 984.061264] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.073045] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332565, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.111696] env[62753]: DEBUG oslo_vmware.api [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Task: {'id': task-1332564, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.239622} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.112436] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 984.112775] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 984.115643] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 984.115643] env[62753]: INFO nova.compute.manager [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Took 1.19 seconds to destroy the instance on the hypervisor. [ 984.115643] env[62753]: DEBUG oslo.service.loopingcall [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.115643] env[62753]: DEBUG nova.compute.manager [-] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 984.115643] env[62753]: DEBUG nova.network.neutron [-] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 984.120584] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "4d20e083-2959-453a-8875-47955bc02613" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 984.202578] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "refresh_cache-91fc519e-c9bb-4605-bdc6-4f463561b988" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 984.204497] env[62753]: DEBUG nova.compute.manager [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Instance network_info: |[{"id": "59ca12e1-f9c6-4e17-a496-248b15733019", "address": "fa:16:3e:3c:35:4c", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59ca12e1-f9", "ovs_interfaceid": "59ca12e1-f9c6-4e17-a496-248b15733019", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 984.206618] env[62753]: DEBUG oslo_concurrency.lockutils [req-6c115c1d-7423-4fc0-a50f-df0779e3de20 req-9cfd48ef-5f3c-4dab-9fff-5accbf5bca78 service nova] Acquired lock "refresh_cache-91fc519e-c9bb-4605-bdc6-4f463561b988" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 984.206618] env[62753]: DEBUG nova.network.neutron [req-6c115c1d-7423-4fc0-a50f-df0779e3de20 req-9cfd48ef-5f3c-4dab-9fff-5accbf5bca78 service nova] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Refreshing network info cache for port 59ca12e1-f9c6-4e17-a496-248b15733019 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 984.207544] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:3c:35:4c', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a10c88d7-d13f-44fd-acee-7a734eb5f56a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '59ca12e1-f9c6-4e17-a496-248b15733019', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 984.216618] env[62753]: DEBUG oslo.service.loopingcall [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 984.223477] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 984.226855] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-31c25bbd-ac73-4dbc-8838-8a4d1c5141ad {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.254034] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 984.254034] env[62753]: value = "task-1332566" [ 984.254034] env[62753]: _type = "Task" [ 984.254034] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.267170] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332566, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.304015] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35fa80b2-3651-44b3-8cbf-ab5e6c9483bb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.313223] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-818ac9a8-3b77-4baa-9cde-2bc1b7f1b710 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.351020] env[62753]: DEBUG oslo_vmware.rw_handles [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e209be-2ca4-dc01-c311-ed193decc7f3/disk-0.vmdk. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 984.351020] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9471d8c6-1e94-424b-9d29-57db43b6aaa7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.354067] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b93c307c-c60e-4b29-8226-272bc8065fa7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.363475] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d0e12cda-e96f-4bb9-acba-12e8e9c84a12 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.367517] env[62753]: DEBUG oslo_vmware.rw_handles [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e209be-2ca4-dc01-c311-ed193decc7f3/disk-0.vmdk is in state: ready. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 984.367855] env[62753]: ERROR oslo_vmware.rw_handles [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e209be-2ca4-dc01-c311-ed193decc7f3/disk-0.vmdk due to incomplete transfer. [ 984.368200] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-1a3bacb1-8b83-4cf7-b8d9-13697b1363d5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.381815] env[62753]: DEBUG nova.compute.provider_tree [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 984.389019] env[62753]: DEBUG oslo_vmware.rw_handles [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52e209be-2ca4-dc01-c311-ed193decc7f3/disk-0.vmdk. {{(pid=62753) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 984.389019] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Uploaded image 3ebd5a0c-fc07-4177-9615-bef143e04e30 to the Glance image server {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 984.390641] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Destroying the VM {{(pid=62753) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 984.391054] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-c2829a90-4227-4d7b-964e-d569a059cf55 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.402018] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 984.402018] env[62753]: value = "task-1332567" [ 984.402018] env[62753]: _type = "Task" [ 984.402018] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.411394] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332567, 'name': Destroy_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.500819] env[62753]: DEBUG oslo_vmware.api [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332561, 'name': PowerOnVM_Task, 'duration_secs': 0.606515} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.501082] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 984.501298] env[62753]: INFO nova.compute.manager [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Took 8.15 seconds to spawn the instance on the hypervisor. [ 984.503027] env[62753]: DEBUG nova.compute.manager [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 984.503027] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e82fc0ae-7a5f-419e-8b01-134b5cd83579 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.574839] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332565, 'name': ReconfigVM_Task, 'duration_secs': 0.168353} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 984.575203] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 984.575498] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6bba65a8-d5aa-43ef-8004-7c7cf57f3476 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 984.584501] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 984.584501] env[62753]: value = "task-1332568" [ 984.584501] env[62753]: _type = "Task" [ 984.584501] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 984.594389] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332568, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.767353] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332566, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.886426] env[62753]: DEBUG nova.scheduler.client.report [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 984.914185] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332567, 'name': Destroy_Task} progress is 33%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 984.963488] env[62753]: DEBUG nova.network.neutron [req-6c115c1d-7423-4fc0-a50f-df0779e3de20 req-9cfd48ef-5f3c-4dab-9fff-5accbf5bca78 service nova] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Updated VIF entry in instance network info cache for port 59ca12e1-f9c6-4e17-a496-248b15733019. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 984.963488] env[62753]: DEBUG nova.network.neutron [req-6c115c1d-7423-4fc0-a50f-df0779e3de20 req-9cfd48ef-5f3c-4dab-9fff-5accbf5bca78 service nova] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Updating instance_info_cache with network_info: [{"id": "59ca12e1-f9c6-4e17-a496-248b15733019", "address": "fa:16:3e:3c:35:4c", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap59ca12e1-f9", "ovs_interfaceid": "59ca12e1-f9c6-4e17-a496-248b15733019", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.026848] env[62753]: INFO nova.compute.manager [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Took 18.46 seconds to build instance. [ 985.050696] env[62753]: DEBUG nova.compute.manager [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 985.074073] env[62753]: DEBUG nova.virt.hardware [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 985.074459] env[62753]: DEBUG nova.virt.hardware [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 985.074891] env[62753]: DEBUG nova.virt.hardware [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 985.075185] env[62753]: DEBUG nova.virt.hardware [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 985.075185] env[62753]: DEBUG nova.virt.hardware [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 985.075414] env[62753]: DEBUG nova.virt.hardware [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 985.075691] env[62753]: DEBUG nova.virt.hardware [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 985.075920] env[62753]: DEBUG nova.virt.hardware [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 985.076189] env[62753]: DEBUG nova.virt.hardware [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 985.076435] env[62753]: DEBUG nova.virt.hardware [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 985.076687] env[62753]: DEBUG nova.virt.hardware [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 985.077871] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4f7e92d1-42e2-4c9b-b729-fce466abe26d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.093123] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c7ed5d1-9c37-49a6-8150-2174dce31429 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.101559] env[62753]: DEBUG oslo_vmware.api [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332568, 'name': PowerOnVM_Task, 'duration_secs': 0.416679} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.111124] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 985.114445] env[62753]: DEBUG nova.compute.manager [None req-1eb2502e-f80d-42e9-9d6d-365a295ecd98 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 985.115761] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de0d4e6d-a748-4119-a7a9-ab19572e6cb7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.176487] env[62753]: DEBUG nova.network.neutron [-] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 985.269316] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332566, 'name': CreateVM_Task, 'duration_secs': 0.697109} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.269515] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 985.270218] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.270391] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.270715] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 985.271472] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca930e94-5c68-430b-8e5d-e3f65a743bdc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.276506] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 985.276506] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]524e1f9d-36b7-5dab-ea23-c565effa894f" [ 985.276506] env[62753]: _type = "Task" [ 985.276506] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.285064] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]524e1f9d-36b7-5dab-ea23-c565effa894f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.392776] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.374s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.393360] env[62753]: DEBUG nova.compute.manager [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 985.396022] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.877s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 985.396262] env[62753]: DEBUG nova.objects.instance [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lazy-loading 'resources' on Instance uuid b5674964-1928-4ecb-b1fd-8f60a94b4270 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 985.410912] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332567, 'name': Destroy_Task, 'duration_secs': 0.98775} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.411261] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Destroyed the VM [ 985.411465] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Deleting Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 985.411759] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-7c5eb0d3-1413-4571-a76d-5ba643d589f6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.419576] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 985.419576] env[62753]: value = "task-1332569" [ 985.419576] env[62753]: _type = "Task" [ 985.419576] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.428254] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332569, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.453092] env[62753]: DEBUG nova.compute.manager [req-cc4fda25-900e-42eb-b378-2aa33750b64d req-57715583-6496-47a6-9eb2-664264a7adf5 service nova] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Received event network-vif-deleted-4037b488-2f26-40ee-b044-3b4fec79a454 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 985.464906] env[62753]: DEBUG oslo_concurrency.lockutils [req-6c115c1d-7423-4fc0-a50f-df0779e3de20 req-9cfd48ef-5f3c-4dab-9fff-5accbf5bca78 service nova] Releasing lock "refresh_cache-91fc519e-c9bb-4605-bdc6-4f463561b988" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.528422] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9b7b7e8e-4ae1-4468-8dac-a027d4d93e70 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 19.975s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 985.573271] env[62753]: DEBUG nova.network.neutron [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Successfully updated port: 011eae16-b648-4e2e-abfb-e424e05e41c2 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 985.679375] env[62753]: INFO nova.compute.manager [-] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Took 1.57 seconds to deallocate network for instance. [ 985.788393] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]524e1f9d-36b7-5dab-ea23-c565effa894f, 'name': SearchDatastore_Task, 'duration_secs': 0.029559} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 985.788944] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 985.789265] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 985.789522] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 985.789681] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 985.789870] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 985.790959] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-53a1d842-2f76-4f8f-a594-c1ceb9223bb3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.800483] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 985.800483] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 985.801018] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7cae5733-9627-4b99-894e-59edcfe8dd0f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 985.807772] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 985.807772] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f92b14-7d2c-a301-5a77-4d90abf5c00b" [ 985.807772] env[62753]: _type = "Task" [ 985.807772] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 985.816705] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f92b14-7d2c-a301-5a77-4d90abf5c00b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.898887] env[62753]: DEBUG nova.compute.utils [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 985.900307] env[62753]: DEBUG nova.compute.manager [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 985.900450] env[62753]: DEBUG nova.network.neutron [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 985.930702] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332569, 'name': RemoveSnapshot_Task} progress is 16%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 985.968528] env[62753]: DEBUG nova.policy [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'cbc9dc744e144ff28cf4c44fb043094f', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e376836411ae4854965adf2923eaaffd', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 986.077612] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "refresh_cache-04fd0022-f20a-4217-9e47-5381635b17a6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 986.077612] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquired lock "refresh_cache-04fd0022-f20a-4217-9e47-5381635b17a6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 986.077612] env[62753]: DEBUG nova.network.neutron [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 986.127616] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-385b4d17-a17f-4be4-8ecd-c7476f0d1c80 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.142020] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466cdf25-6020-4f40-9a4e-23bd4ef698e3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.168811] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9da40f91-6347-4a29-8a1f-113715adfd33 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.176199] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6eefef3f-d714-47c6-8fc6-3cda4f253f4d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.190657] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.190717] env[62753]: DEBUG nova.compute.provider_tree [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 986.282268] env[62753]: DEBUG nova.network.neutron [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Successfully created port: 99cbc723-6896-4ee7-bc42-e8fdac96ea65 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 986.318587] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f92b14-7d2c-a301-5a77-4d90abf5c00b, 'name': SearchDatastore_Task, 'duration_secs': 0.01108} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.319378] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4ab92942-f0b2-48ab-8c51-f08fc8f59f75 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.325273] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 986.325273] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529e4407-f6c2-c6e4-af52-d02a788795bb" [ 986.325273] env[62753]: _type = "Task" [ 986.325273] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.334107] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529e4407-f6c2-c6e4-af52-d02a788795bb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.404305] env[62753]: DEBUG nova.compute.manager [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 986.430386] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332569, 'name': RemoveSnapshot_Task} progress is 16%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.572775] env[62753]: DEBUG nova.compute.manager [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Stashing vm_state: active {{(pid=62753) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 986.616239] env[62753]: DEBUG nova.network.neutron [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 986.696869] env[62753]: DEBUG nova.scheduler.client.report [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 986.700255] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "9d7066e1-3721-42d2-8e80-91eacbbd8b80" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 986.700535] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "9d7066e1-3721-42d2-8e80-91eacbbd8b80" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 986.761116] env[62753]: DEBUG nova.network.neutron [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Updating instance_info_cache with network_info: [{"id": "011eae16-b648-4e2e-abfb-e424e05e41c2", "address": "fa:16:3e:32:28:cb", "network": {"id": "61c6e640-d2e6-4b3b-b191-7da9249dd339", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-703836479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8925b84dcf9a47fbaf2eb8044b3850fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap011eae16-b6", "ovs_interfaceid": "011eae16-b648-4e2e-abfb-e424e05e41c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 986.836314] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]529e4407-f6c2-c6e4-af52-d02a788795bb, 'name': SearchDatastore_Task, 'duration_secs': 0.011847} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.836741] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 986.837055] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 91fc519e-c9bb-4605-bdc6-4f463561b988/91fc519e-c9bb-4605-bdc6-4f463561b988.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 986.837366] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-680b2cc0-de74-4a9a-9fc2-aff58ded6051 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 986.844451] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 986.844451] env[62753]: value = "task-1332570" [ 986.844451] env[62753]: _type = "Task" [ 986.844451] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 986.852842] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332570, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 986.930473] env[62753]: DEBUG oslo_vmware.api [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332569, 'name': RemoveSnapshot_Task, 'duration_secs': 1.057399} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 986.930861] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Deleted Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 986.931229] env[62753]: INFO nova.compute.manager [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Took 17.91 seconds to snapshot the instance on the hypervisor. [ 987.097309] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.203357] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.807s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.206103] env[62753]: DEBUG nova.compute.manager [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 987.209787] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.833s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.210149] env[62753]: DEBUG nova.objects.instance [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lazy-loading 'resources' on Instance uuid 4d20e083-2959-453a-8875-47955bc02613 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.231233] env[62753]: INFO nova.scheduler.client.report [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Deleted allocations for instance b5674964-1928-4ecb-b1fd-8f60a94b4270 [ 987.263723] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Releasing lock "refresh_cache-04fd0022-f20a-4217-9e47-5381635b17a6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 987.264231] env[62753]: DEBUG nova.compute.manager [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Instance network_info: |[{"id": "011eae16-b648-4e2e-abfb-e424e05e41c2", "address": "fa:16:3e:32:28:cb", "network": {"id": "61c6e640-d2e6-4b3b-b191-7da9249dd339", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-703836479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8925b84dcf9a47fbaf2eb8044b3850fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap011eae16-b6", "ovs_interfaceid": "011eae16-b648-4e2e-abfb-e424e05e41c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 987.264780] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:32:28:cb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '510d3c47-3615-43d5-aa5d-a279fd915e71', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '011eae16-b648-4e2e-abfb-e424e05e41c2', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 987.274162] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Creating folder: Project (8925b84dcf9a47fbaf2eb8044b3850fa). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 987.274713] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b0705f80-fb94-4c15-95a8-e7902752e597 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.286480] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Created folder: Project (8925b84dcf9a47fbaf2eb8044b3850fa) in parent group-v284541. [ 987.286733] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Creating folder: Instances. Parent ref: group-v284652. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 987.287020] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-b3495a86-a7b0-44be-b571-424254f88f07 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.297088] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Created folder: Instances in parent group-v284652. [ 987.297391] env[62753]: DEBUG oslo.service.loopingcall [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 987.297588] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 987.297811] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-b2d6ca69-bfa9-42b1-97c4-a2f5539f68bd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.318542] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 987.318542] env[62753]: value = "task-1332573" [ 987.318542] env[62753]: _type = "Task" [ 987.318542] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.327228] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332573, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.354079] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332570, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.390585] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "interface-c3679c66-2763-4a04-8d14-e103bd952798-2498c430-c268-4e1e-b331-0a3b05eeae0a" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.390950] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "interface-c3679c66-2763-4a04-8d14-e103bd952798-2498c430-c268-4e1e-b331-0a3b05eeae0a" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.391369] env[62753]: DEBUG nova.objects.instance [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lazy-loading 'flavor' on Instance uuid c3679c66-2763-4a04-8d14-e103bd952798 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.414843] env[62753]: DEBUG nova.compute.manager [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 987.447761] env[62753]: DEBUG nova.virt.hardware [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 987.448032] env[62753]: DEBUG nova.virt.hardware [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 987.448202] env[62753]: DEBUG nova.virt.hardware [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 987.448396] env[62753]: DEBUG nova.virt.hardware [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 987.448570] env[62753]: DEBUG nova.virt.hardware [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 987.448727] env[62753]: DEBUG nova.virt.hardware [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 987.448942] env[62753]: DEBUG nova.virt.hardware [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 987.449153] env[62753]: DEBUG nova.virt.hardware [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 987.449400] env[62753]: DEBUG nova.virt.hardware [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 987.449600] env[62753]: DEBUG nova.virt.hardware [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 987.449805] env[62753]: DEBUG nova.virt.hardware [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 987.450863] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-09cf8334-53a5-43dd-b331-957b36c25bec {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.459478] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0c909de5-b337-4600-a3e8-f0d2bbc2bc05 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.481610] env[62753]: DEBUG nova.compute.manager [None req-3d3e4b2f-f259-4389-9bf7-e87f98bddbc9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Found 2 images (rotation: 2) {{(pid=62753) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 987.521093] env[62753]: DEBUG nova.compute.manager [req-6f687afb-75bd-4a71-8608-cef7da39299b req-c9bf26f0-191c-41d0-a73d-c3f3c59ada0e service nova] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Received event network-vif-plugged-011eae16-b648-4e2e-abfb-e424e05e41c2 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.521529] env[62753]: DEBUG oslo_concurrency.lockutils [req-6f687afb-75bd-4a71-8608-cef7da39299b req-c9bf26f0-191c-41d0-a73d-c3f3c59ada0e service nova] Acquiring lock "04fd0022-f20a-4217-9e47-5381635b17a6-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.521769] env[62753]: DEBUG oslo_concurrency.lockutils [req-6f687afb-75bd-4a71-8608-cef7da39299b req-c9bf26f0-191c-41d0-a73d-c3f3c59ada0e service nova] Lock "04fd0022-f20a-4217-9e47-5381635b17a6-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 987.521898] env[62753]: DEBUG oslo_concurrency.lockutils [req-6f687afb-75bd-4a71-8608-cef7da39299b req-c9bf26f0-191c-41d0-a73d-c3f3c59ada0e service nova] Lock "04fd0022-f20a-4217-9e47-5381635b17a6-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.522144] env[62753]: DEBUG nova.compute.manager [req-6f687afb-75bd-4a71-8608-cef7da39299b req-c9bf26f0-191c-41d0-a73d-c3f3c59ada0e service nova] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] No waiting events found dispatching network-vif-plugged-011eae16-b648-4e2e-abfb-e424e05e41c2 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 987.522396] env[62753]: WARNING nova.compute.manager [req-6f687afb-75bd-4a71-8608-cef7da39299b req-c9bf26f0-191c-41d0-a73d-c3f3c59ada0e service nova] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Received unexpected event network-vif-plugged-011eae16-b648-4e2e-abfb-e424e05e41c2 for instance with vm_state building and task_state spawning. [ 987.522547] env[62753]: DEBUG nova.compute.manager [req-6f687afb-75bd-4a71-8608-cef7da39299b req-c9bf26f0-191c-41d0-a73d-c3f3c59ada0e service nova] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Received event network-changed-011eae16-b648-4e2e-abfb-e424e05e41c2 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 987.522675] env[62753]: DEBUG nova.compute.manager [req-6f687afb-75bd-4a71-8608-cef7da39299b req-c9bf26f0-191c-41d0-a73d-c3f3c59ada0e service nova] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Refreshing instance network info cache due to event network-changed-011eae16-b648-4e2e-abfb-e424e05e41c2. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 987.522886] env[62753]: DEBUG oslo_concurrency.lockutils [req-6f687afb-75bd-4a71-8608-cef7da39299b req-c9bf26f0-191c-41d0-a73d-c3f3c59ada0e service nova] Acquiring lock "refresh_cache-04fd0022-f20a-4217-9e47-5381635b17a6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.523063] env[62753]: DEBUG oslo_concurrency.lockutils [req-6f687afb-75bd-4a71-8608-cef7da39299b req-c9bf26f0-191c-41d0-a73d-c3f3c59ada0e service nova] Acquired lock "refresh_cache-04fd0022-f20a-4217-9e47-5381635b17a6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.523245] env[62753]: DEBUG nova.network.neutron [req-6f687afb-75bd-4a71-8608-cef7da39299b req-c9bf26f0-191c-41d0-a73d-c3f3c59ada0e service nova] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Refreshing network info cache for port 011eae16-b648-4e2e-abfb-e424e05e41c2 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 987.717760] env[62753]: DEBUG nova.objects.instance [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lazy-loading 'numa_topology' on Instance uuid 4d20e083-2959-453a-8875-47955bc02613 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 987.731674] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 987.744866] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e5def9a9-7869-41f0-adfa-88c476fe44d6 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "b5674964-1928-4ecb-b1fd-8f60a94b4270" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.341s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 987.746104] env[62753]: DEBUG nova.network.neutron [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Successfully updated port: 99cbc723-6896-4ee7-bc42-e8fdac96ea65 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 987.829239] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332573, 'name': CreateVM_Task, 'duration_secs': 0.383496} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.829484] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 987.830183] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 987.830357] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 987.830733] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 987.830981] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dfe4a230-9579-4887-9b5b-ce7f99327ef8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.835456] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 987.835456] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5298a074-434c-8c38-32c3-d097d931bdeb" [ 987.835456] env[62753]: _type = "Task" [ 987.835456] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.843224] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5298a074-434c-8c38-32c3-d097d931bdeb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 987.853693] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332570, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.56165} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 987.853976] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 91fc519e-c9bb-4605-bdc6-4f463561b988/91fc519e-c9bb-4605-bdc6-4f463561b988.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 987.854226] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 987.854481] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2f86cd5d-989e-4f1f-944d-b9e408ffaa23 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 987.861214] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 987.861214] env[62753]: value = "task-1332574" [ 987.861214] env[62753]: _type = "Task" [ 987.861214] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 987.868634] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332574, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.006559] env[62753]: DEBUG nova.objects.instance [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lazy-loading 'pci_requests' on Instance uuid c3679c66-2763-4a04-8d14-e103bd952798 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.220461] env[62753]: DEBUG nova.objects.base [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Object Instance<4d20e083-2959-453a-8875-47955bc02613> lazy-loaded attributes: resources,numa_topology {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 988.251562] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "refresh_cache-e1c59139-ff02-453e-b54f-a56b1e40fdb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.251705] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired lock "refresh_cache-e1c59139-ff02-453e-b54f-a56b1e40fdb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.251858] env[62753]: DEBUG nova.network.neutron [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 988.255891] env[62753]: DEBUG nova.network.neutron [req-6f687afb-75bd-4a71-8608-cef7da39299b req-c9bf26f0-191c-41d0-a73d-c3f3c59ada0e service nova] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Updated VIF entry in instance network info cache for port 011eae16-b648-4e2e-abfb-e424e05e41c2. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 988.256253] env[62753]: DEBUG nova.network.neutron [req-6f687afb-75bd-4a71-8608-cef7da39299b req-c9bf26f0-191c-41d0-a73d-c3f3c59ada0e service nova] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Updating instance_info_cache with network_info: [{"id": "011eae16-b648-4e2e-abfb-e424e05e41c2", "address": "fa:16:3e:32:28:cb", "network": {"id": "61c6e640-d2e6-4b3b-b191-7da9249dd339", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-703836479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8925b84dcf9a47fbaf2eb8044b3850fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap011eae16-b6", "ovs_interfaceid": "011eae16-b648-4e2e-abfb-e424e05e41c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.346145] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5298a074-434c-8c38-32c3-d097d931bdeb, 'name': SearchDatastore_Task, 'duration_secs': 0.01597} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.346459] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.346691] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 988.346922] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 988.347087] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 988.347271] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 988.347529] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3769c0a3-98c2-4808-a104-b2ff94c06296 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.361032] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 988.361236] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 988.362043] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f6c44000-9df1-4996-a18e-35875601ef95 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.373824] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332574, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.080663} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.374954] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 988.375319] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 988.375319] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]520fdb93-abc9-b39c-e016-ab59140c476b" [ 988.375319] env[62753]: _type = "Task" [ 988.375319] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.375988] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f09f5ca8-e031-44fd-93af-35c8abde6ec0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.387992] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]520fdb93-abc9-b39c-e016-ab59140c476b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.405900] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Reconfiguring VM instance instance-00000057 to attach disk [datastore2] 91fc519e-c9bb-4605-bdc6-4f463561b988/91fc519e-c9bb-4605-bdc6-4f463561b988.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 988.408937] env[62753]: DEBUG nova.compute.manager [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 988.409217] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-52b534be-0958-4a43-b6c5-d5bed4f9b0c4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.423817] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78d4ad2-6e8d-4ddb-a1f4-a8ddb671c857 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.434827] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 988.434827] env[62753]: value = "task-1332575" [ 988.434827] env[62753]: _type = "Task" [ 988.434827] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.435525] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4667430d-5a67-4487-a2b9-79eb045f16f1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.447544] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f47b605b-b24f-4f5d-8ed9-cd4d3d59c2a4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.450999] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332575, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.480353] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30b5c56f-f9b9-40ac-becf-47abbfbd1e6e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.488445] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-034cea91-1fdc-407d-9876-f89edf4887ab {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.505053] env[62753]: DEBUG nova.compute.provider_tree [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 988.509025] env[62753]: DEBUG nova.objects.base [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Object Instance lazy-loaded attributes: flavor,pci_requests {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 988.509154] env[62753]: DEBUG nova.network.neutron [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 988.573057] env[62753]: DEBUG nova.policy [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9931b40cfd7846038805c6a4caedaac0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b8e81660b30408c998e412f5fa81469', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 988.760809] env[62753]: DEBUG oslo_concurrency.lockutils [req-6f687afb-75bd-4a71-8608-cef7da39299b req-c9bf26f0-191c-41d0-a73d-c3f3c59ada0e service nova] Releasing lock "refresh_cache-04fd0022-f20a-4217-9e47-5381635b17a6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 988.794125] env[62753]: DEBUG nova.network.neutron [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 988.889064] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]520fdb93-abc9-b39c-e016-ab59140c476b, 'name': SearchDatastore_Task, 'duration_secs': 0.060507} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.889835] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-af4cfbd0-e4da-4a36-a301-3ff8d8ce7343 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.894696] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 988.894696] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525eccb6-f6c2-56a0-470e-96daa8c35e75" [ 988.894696] env[62753]: _type = "Task" [ 988.894696] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.902196] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525eccb6-f6c2-56a0-470e-96daa8c35e75, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 988.934729] env[62753]: INFO nova.compute.manager [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] instance snapshotting [ 988.935662] env[62753]: DEBUG nova.objects.instance [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lazy-loading 'flavor' on Instance uuid c73b1ae0-dc98-47f7-babf-e96169384785 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 988.939580] env[62753]: DEBUG nova.network.neutron [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Updating instance_info_cache with network_info: [{"id": "99cbc723-6896-4ee7-bc42-e8fdac96ea65", "address": "fa:16:3e:2a:df:10", "network": {"id": "8c5edc3b-7faf-47f2-9736-7093f9ff72de", "bridge": "br-int", "label": "tempest-ImagesTestJSON-588052284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e376836411ae4854965adf2923eaaffd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76f377cd-5966-49b4-9210-907f592c694e", "external-id": "nsx-vlan-transportzone-124", "segmentation_id": 124, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99cbc723-68", "ovs_interfaceid": "99cbc723-6896-4ee7-bc42-e8fdac96ea65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 988.950493] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332575, 'name': ReconfigVM_Task, 'duration_secs': 0.324821} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 988.950761] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Reconfigured VM instance instance-00000057 to attach disk [datastore2] 91fc519e-c9bb-4605-bdc6-4f463561b988/91fc519e-c9bb-4605-bdc6-4f463561b988.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 988.951634] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7d9a8577-3047-4c63-a833-794e87a99855 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 988.959532] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 988.959532] env[62753]: value = "task-1332576" [ 988.959532] env[62753]: _type = "Task" [ 988.959532] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 988.967621] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332576, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.006427] env[62753]: DEBUG nova.scheduler.client.report [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 989.404924] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525eccb6-f6c2-56a0-470e-96daa8c35e75, 'name': SearchDatastore_Task, 'duration_secs': 0.014668} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.405220] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.405460] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 04fd0022-f20a-4217-9e47-5381635b17a6/04fd0022-f20a-4217-9e47-5381635b17a6.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 989.405718] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-22d0e362-5454-4939-9e6e-4758bbde4270 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.411796] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 989.411796] env[62753]: value = "task-1332577" [ 989.411796] env[62753]: _type = "Task" [ 989.411796] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.419171] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332577, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.441402] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-466e79e5-0aad-4e01-acf4-3bc8d09a509c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.461210] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Releasing lock "refresh_cache-e1c59139-ff02-453e-b54f-a56b1e40fdb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 989.461515] env[62753]: DEBUG nova.compute.manager [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Instance network_info: |[{"id": "99cbc723-6896-4ee7-bc42-e8fdac96ea65", "address": "fa:16:3e:2a:df:10", "network": {"id": "8c5edc3b-7faf-47f2-9736-7093f9ff72de", "bridge": "br-int", "label": "tempest-ImagesTestJSON-588052284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e376836411ae4854965adf2923eaaffd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76f377cd-5966-49b4-9210-907f592c694e", "external-id": "nsx-vlan-transportzone-124", "segmentation_id": 124, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99cbc723-68", "ovs_interfaceid": "99cbc723-6896-4ee7-bc42-e8fdac96ea65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 989.462117] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:2a:df:10', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '76f377cd-5966-49b4-9210-907f592c694e', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '99cbc723-6896-4ee7-bc42-e8fdac96ea65', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 989.469867] env[62753]: DEBUG oslo.service.loopingcall [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 989.473374] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-667ff748-8814-4b5f-a365-59f5b5f2c9a3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.475990] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 989.476227] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-40986911-61f4-4450-8c48-da962bf0f930 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.496439] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332576, 'name': Rename_Task, 'duration_secs': 0.166294} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.499051] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 989.499282] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 989.499282] env[62753]: value = "task-1332578" [ 989.499282] env[62753]: _type = "Task" [ 989.499282] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.502120] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-b0d48fdc-2455-47b0-a595-d1a275f28dbf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.512588] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.303s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.515375] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332578, 'name': CreateVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.516710] env[62753]: DEBUG oslo_concurrency.lockutils [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.793s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.518121] env[62753]: INFO nova.compute.claims [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 989.520769] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 989.520769] env[62753]: value = "task-1332579" [ 989.520769] env[62753]: _type = "Task" [ 989.520769] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.530116] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332579, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 989.629252] env[62753]: DEBUG nova.compute.manager [req-3f1bfe12-4cde-4bef-aee7-b50b443fa165 req-3baf2004-5237-4667-848c-c5f7631fd790 service nova] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Received event network-vif-plugged-99cbc723-6896-4ee7-bc42-e8fdac96ea65 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.629647] env[62753]: DEBUG oslo_concurrency.lockutils [req-3f1bfe12-4cde-4bef-aee7-b50b443fa165 req-3baf2004-5237-4667-848c-c5f7631fd790 service nova] Acquiring lock "e1c59139-ff02-453e-b54f-a56b1e40fdb5-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 989.629915] env[62753]: DEBUG oslo_concurrency.lockutils [req-3f1bfe12-4cde-4bef-aee7-b50b443fa165 req-3baf2004-5237-4667-848c-c5f7631fd790 service nova] Lock "e1c59139-ff02-453e-b54f-a56b1e40fdb5-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 989.630113] env[62753]: DEBUG oslo_concurrency.lockutils [req-3f1bfe12-4cde-4bef-aee7-b50b443fa165 req-3baf2004-5237-4667-848c-c5f7631fd790 service nova] Lock "e1c59139-ff02-453e-b54f-a56b1e40fdb5-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 989.630332] env[62753]: DEBUG nova.compute.manager [req-3f1bfe12-4cde-4bef-aee7-b50b443fa165 req-3baf2004-5237-4667-848c-c5f7631fd790 service nova] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] No waiting events found dispatching network-vif-plugged-99cbc723-6896-4ee7-bc42-e8fdac96ea65 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 989.630554] env[62753]: WARNING nova.compute.manager [req-3f1bfe12-4cde-4bef-aee7-b50b443fa165 req-3baf2004-5237-4667-848c-c5f7631fd790 service nova] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Received unexpected event network-vif-plugged-99cbc723-6896-4ee7-bc42-e8fdac96ea65 for instance with vm_state building and task_state spawning. [ 989.630783] env[62753]: DEBUG nova.compute.manager [req-3f1bfe12-4cde-4bef-aee7-b50b443fa165 req-3baf2004-5237-4667-848c-c5f7631fd790 service nova] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Received event network-changed-99cbc723-6896-4ee7-bc42-e8fdac96ea65 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 989.631007] env[62753]: DEBUG nova.compute.manager [req-3f1bfe12-4cde-4bef-aee7-b50b443fa165 req-3baf2004-5237-4667-848c-c5f7631fd790 service nova] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Refreshing instance network info cache due to event network-changed-99cbc723-6896-4ee7-bc42-e8fdac96ea65. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 989.631260] env[62753]: DEBUG oslo_concurrency.lockutils [req-3f1bfe12-4cde-4bef-aee7-b50b443fa165 req-3baf2004-5237-4667-848c-c5f7631fd790 service nova] Acquiring lock "refresh_cache-e1c59139-ff02-453e-b54f-a56b1e40fdb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 989.631465] env[62753]: DEBUG oslo_concurrency.lockutils [req-3f1bfe12-4cde-4bef-aee7-b50b443fa165 req-3baf2004-5237-4667-848c-c5f7631fd790 service nova] Acquired lock "refresh_cache-e1c59139-ff02-453e-b54f-a56b1e40fdb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 989.631687] env[62753]: DEBUG nova.network.neutron [req-3f1bfe12-4cde-4bef-aee7-b50b443fa165 req-3baf2004-5237-4667-848c-c5f7631fd790 service nova] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Refreshing network info cache for port 99cbc723-6896-4ee7-bc42-e8fdac96ea65 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 989.923985] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332577, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.488125} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 989.924396] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 04fd0022-f20a-4217-9e47-5381635b17a6/04fd0022-f20a-4217-9e47-5381635b17a6.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 989.924488] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 989.924754] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-a4a31d68-dddc-46f3-91d7-030490b8ddb5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 989.947134] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 989.947134] env[62753]: value = "task-1332580" [ 989.947134] env[62753]: _type = "Task" [ 989.947134] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 989.957503] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332580, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.004307] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Creating Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 990.004680] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-d55a856f-6b60-4139-bb81-5aea93e55db2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.014931] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332578, 'name': CreateVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.017833] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 990.017833] env[62753]: value = "task-1332581" [ 990.017833] env[62753]: _type = "Task" [ 990.017833] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.024072] env[62753]: DEBUG oslo_concurrency.lockutils [None req-f4ecfcf2-3cb9-400d-b7f9-12c26affa3d6 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "4d20e083-2959-453a-8875-47955bc02613" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 30.686s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 990.027695] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "4d20e083-2959-453a-8875-47955bc02613" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 5.907s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 990.028295] env[62753]: INFO nova.compute.manager [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Unshelving [ 990.038873] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332581, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.044859] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332579, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.153163] env[62753]: DEBUG nova.network.neutron [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Successfully updated port: 2498c430-c268-4e1e-b331-0a3b05eeae0a {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 990.351878] env[62753]: DEBUG nova.network.neutron [req-3f1bfe12-4cde-4bef-aee7-b50b443fa165 req-3baf2004-5237-4667-848c-c5f7631fd790 service nova] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Updated VIF entry in instance network info cache for port 99cbc723-6896-4ee7-bc42-e8fdac96ea65. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 990.352419] env[62753]: DEBUG nova.network.neutron [req-3f1bfe12-4cde-4bef-aee7-b50b443fa165 req-3baf2004-5237-4667-848c-c5f7631fd790 service nova] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Updating instance_info_cache with network_info: [{"id": "99cbc723-6896-4ee7-bc42-e8fdac96ea65", "address": "fa:16:3e:2a:df:10", "network": {"id": "8c5edc3b-7faf-47f2-9736-7093f9ff72de", "bridge": "br-int", "label": "tempest-ImagesTestJSON-588052284-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e376836411ae4854965adf2923eaaffd", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "76f377cd-5966-49b4-9210-907f592c694e", "external-id": "nsx-vlan-transportzone-124", "segmentation_id": 124, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap99cbc723-68", "ovs_interfaceid": "99cbc723-6896-4ee7-bc42-e8fdac96ea65", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 990.457063] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332580, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.087608} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.457063] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 990.457864] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-70ae1f5b-65dc-4c7f-8f1b-07883872bc80 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.479468] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Reconfiguring VM instance instance-00000058 to attach disk [datastore2] 04fd0022-f20a-4217-9e47-5381635b17a6/04fd0022-f20a-4217-9e47-5381635b17a6.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 990.479750] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac786793-c1b0-45db-8b21-8cef3893d1a5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.500586] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 990.500586] env[62753]: value = "task-1332582" [ 990.500586] env[62753]: _type = "Task" [ 990.500586] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.511296] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332582, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.516334] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332578, 'name': CreateVM_Task, 'duration_secs': 0.544342} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.516488] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 990.517152] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.517320] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.517641] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 990.517891] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-022a52cd-d8ec-419c-94a0-0b7db1ec69d6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.525040] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 990.525040] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52496a0c-bc9c-ab00-833f-201a95333b26" [ 990.525040] env[62753]: _type = "Task" [ 990.525040] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.537090] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332581, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.549530] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52496a0c-bc9c-ab00-833f-201a95333b26, 'name': SearchDatastore_Task, 'duration_secs': 0.015754} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.553196] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 990.553349] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 990.553610] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.553815] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.554013] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 990.554386] env[62753]: DEBUG oslo_vmware.api [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332579, 'name': PowerOnVM_Task, 'duration_secs': 0.71553} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 990.554657] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0a549b73-1170-4605-90f8-70e93e8aabc4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.557868] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 990.558122] env[62753]: INFO nova.compute.manager [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Took 9.66 seconds to spawn the instance on the hypervisor. [ 990.558319] env[62753]: DEBUG nova.compute.manager [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 990.559222] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41e09f46-195c-457f-bb3a-38669317473e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.576027] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 990.576027] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 990.576614] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b2e93edb-ef68-43ae-9b0e-03f8197dd538 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.585970] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 990.585970] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52cc6b52-30b9-d123-3244-99613ac5a274" [ 990.585970] env[62753]: _type = "Task" [ 990.585970] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 990.595019] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52cc6b52-30b9-d123-3244-99613ac5a274, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 990.656329] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 990.656526] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 990.656708] env[62753]: DEBUG nova.network.neutron [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 990.728510] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1468f299-903c-46fe-ac8c-a58f7e244157 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.736384] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e935b1c-73ce-499a-95ef-8c68a592cc6c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.768581] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa8d9ae3-5471-4b44-9b28-365564d1fba9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.776143] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb759a7b-2685-4710-8eb7-290837934b7c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 990.789698] env[62753]: DEBUG nova.compute.provider_tree [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 990.855608] env[62753]: DEBUG oslo_concurrency.lockutils [req-3f1bfe12-4cde-4bef-aee7-b50b443fa165 req-3baf2004-5237-4667-848c-c5f7631fd790 service nova] Releasing lock "refresh_cache-e1c59139-ff02-453e-b54f-a56b1e40fdb5" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.010560] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332582, 'name': ReconfigVM_Task, 'duration_secs': 0.288413} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.010946] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Reconfigured VM instance instance-00000058 to attach disk [datastore2] 04fd0022-f20a-4217-9e47-5381635b17a6/04fd0022-f20a-4217-9e47-5381635b17a6.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 991.011497] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-36ec6d22-889e-4fa3-9041-f31aaefca224 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.017364] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 991.017364] env[62753]: value = "task-1332583" [ 991.017364] env[62753]: _type = "Task" [ 991.017364] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.026954] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332583, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.031541] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332581, 'name': CreateSnapshot_Task, 'duration_secs': 0.572811} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.031891] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Created Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 991.032593] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c27b0327-a2b2-45db-8561-aa6efe58df0f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.059308] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.084643] env[62753]: INFO nova.compute.manager [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Took 15.74 seconds to build instance. [ 991.098585] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52cc6b52-30b9-d123-3244-99613ac5a274, 'name': SearchDatastore_Task, 'duration_secs': 0.017989} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.099692] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4a66df28-39fb-42cb-82da-0032e68f8132 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.106089] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 991.106089] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d52824-c911-36eb-a03d-20dff14c8a79" [ 991.106089] env[62753]: _type = "Task" [ 991.106089] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.114767] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d52824-c911-36eb-a03d-20dff14c8a79, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.189750] env[62753]: WARNING nova.network.neutron [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] 8a9463a1-a861-4ade-8483-640392be0a89 already exists in list: networks containing: ['8a9463a1-a861-4ade-8483-640392be0a89']. ignoring it [ 991.292733] env[62753]: DEBUG nova.scheduler.client.report [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 991.469443] env[62753]: DEBUG nova.network.neutron [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Updating instance_info_cache with network_info: [{"id": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "address": "fa:16:3e:57:5d:75", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12ca481a-0c", "ovs_interfaceid": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2498c430-c268-4e1e-b331-0a3b05eeae0a", "address": "fa:16:3e:f3:a1:e8", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2498c430-c2", "ovs_interfaceid": "2498c430-c268-4e1e-b331-0a3b05eeae0a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 991.527537] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332583, 'name': Rename_Task, 'duration_secs': 0.148847} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.527830] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 991.528129] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4fa8f994-949d-4280-b1d3-66f7ee29d55a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.534289] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 991.534289] env[62753]: value = "task-1332584" [ 991.534289] env[62753]: _type = "Task" [ 991.534289] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.548072] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Creating linked-clone VM from snapshot {{(pid=62753) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 991.548384] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332584, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.548901] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-0ef33b66-c8b4-45db-8cfc-6de6ee1e105c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.556221] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 991.556221] env[62753]: value = "task-1332585" [ 991.556221] env[62753]: _type = "Task" [ 991.556221] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.564512] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332585, 'name': CloneVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.586315] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4af1f343-a492-4654-8ea2-48a1f991a1d4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "91fc519e-c9bb-4605-bdc6-4f463561b988" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.246s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.615767] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d52824-c911-36eb-a03d-20dff14c8a79, 'name': SearchDatastore_Task, 'duration_secs': 0.028435} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 991.616042] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.616315] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] e1c59139-ff02-453e-b54f-a56b1e40fdb5/e1c59139-ff02-453e-b54f-a56b1e40fdb5.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 991.616579] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2fb05cb9-f5b9-4abb-97c8-028de1193477 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.622746] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 991.622746] env[62753]: value = "task-1332586" [ 991.622746] env[62753]: _type = "Task" [ 991.622746] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 991.630653] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332586, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 991.738880] env[62753]: DEBUG nova.compute.manager [req-3ccdbf2a-b459-4d12-aa26-f0d06167cb44 req-7a622bd5-f6d8-4d75-9cca-741035f172a7 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Received event network-vif-plugged-2498c430-c268-4e1e-b331-0a3b05eeae0a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.739115] env[62753]: DEBUG oslo_concurrency.lockutils [req-3ccdbf2a-b459-4d12-aa26-f0d06167cb44 req-7a622bd5-f6d8-4d75-9cca-741035f172a7 service nova] Acquiring lock "c3679c66-2763-4a04-8d14-e103bd952798-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 991.739328] env[62753]: DEBUG oslo_concurrency.lockutils [req-3ccdbf2a-b459-4d12-aa26-f0d06167cb44 req-7a622bd5-f6d8-4d75-9cca-741035f172a7 service nova] Lock "c3679c66-2763-4a04-8d14-e103bd952798-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.739864] env[62753]: DEBUG oslo_concurrency.lockutils [req-3ccdbf2a-b459-4d12-aa26-f0d06167cb44 req-7a622bd5-f6d8-4d75-9cca-741035f172a7 service nova] Lock "c3679c66-2763-4a04-8d14-e103bd952798-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.739864] env[62753]: DEBUG nova.compute.manager [req-3ccdbf2a-b459-4d12-aa26-f0d06167cb44 req-7a622bd5-f6d8-4d75-9cca-741035f172a7 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] No waiting events found dispatching network-vif-plugged-2498c430-c268-4e1e-b331-0a3b05eeae0a {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 991.739864] env[62753]: WARNING nova.compute.manager [req-3ccdbf2a-b459-4d12-aa26-f0d06167cb44 req-7a622bd5-f6d8-4d75-9cca-741035f172a7 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Received unexpected event network-vif-plugged-2498c430-c268-4e1e-b331-0a3b05eeae0a for instance with vm_state active and task_state None. [ 991.740049] env[62753]: DEBUG nova.compute.manager [req-3ccdbf2a-b459-4d12-aa26-f0d06167cb44 req-7a622bd5-f6d8-4d75-9cca-741035f172a7 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Received event network-changed-2498c430-c268-4e1e-b331-0a3b05eeae0a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 991.740120] env[62753]: DEBUG nova.compute.manager [req-3ccdbf2a-b459-4d12-aa26-f0d06167cb44 req-7a622bd5-f6d8-4d75-9cca-741035f172a7 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Refreshing instance network info cache due to event network-changed-2498c430-c268-4e1e-b331-0a3b05eeae0a. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 991.740328] env[62753]: DEBUG oslo_concurrency.lockutils [req-3ccdbf2a-b459-4d12-aa26-f0d06167cb44 req-7a622bd5-f6d8-4d75-9cca-741035f172a7 service nova] Acquiring lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.797961] env[62753]: DEBUG oslo_concurrency.lockutils [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.281s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 991.798394] env[62753]: DEBUG nova.compute.manager [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 991.802140] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.612s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 991.802317] env[62753]: DEBUG nova.objects.instance [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lazy-loading 'resources' on Instance uuid 77cdd901-cc96-4a6f-a696-a65f54a96b1a {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 991.972555] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 991.973336] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 991.973507] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.973840] env[62753]: DEBUG oslo_concurrency.lockutils [req-3ccdbf2a-b459-4d12-aa26-f0d06167cb44 req-7a622bd5-f6d8-4d75-9cca-741035f172a7 service nova] Acquired lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 991.974133] env[62753]: DEBUG nova.network.neutron [req-3ccdbf2a-b459-4d12-aa26-f0d06167cb44 req-7a622bd5-f6d8-4d75-9cca-741035f172a7 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Refreshing network info cache for port 2498c430-c268-4e1e-b331-0a3b05eeae0a {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 991.976324] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3770a99-1dfc-4d1a-b592-e9476293b57f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 991.994791] env[62753]: DEBUG nova.virt.hardware [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 991.995117] env[62753]: DEBUG nova.virt.hardware [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 991.995314] env[62753]: DEBUG nova.virt.hardware [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 991.995572] env[62753]: DEBUG nova.virt.hardware [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 991.995666] env[62753]: DEBUG nova.virt.hardware [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 991.995811] env[62753]: DEBUG nova.virt.hardware [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 991.996131] env[62753]: DEBUG nova.virt.hardware [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 991.996320] env[62753]: DEBUG nova.virt.hardware [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 991.996477] env[62753]: DEBUG nova.virt.hardware [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 991.996655] env[62753]: DEBUG nova.virt.hardware [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 991.996840] env[62753]: DEBUG nova.virt.hardware [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 992.003497] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Reconfiguring VM to attach interface {{(pid=62753) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 992.004947] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5f554cea-9cd1-49f4-9079-60b293702717 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.022926] env[62753]: DEBUG oslo_vmware.api [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 992.022926] env[62753]: value = "task-1332587" [ 992.022926] env[62753]: _type = "Task" [ 992.022926] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.034196] env[62753]: DEBUG oslo_vmware.api [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332587, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.045147] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332584, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.066062] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332585, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.133873] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332586, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.221563] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3556fdab-c30c-4c08-ab9e-ee202a858cb6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "91fc519e-c9bb-4605-bdc6-4f463561b988" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 992.221985] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3556fdab-c30c-4c08-ab9e-ee202a858cb6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "91fc519e-c9bb-4605-bdc6-4f463561b988" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 992.222328] env[62753]: DEBUG nova.compute.manager [None req-3556fdab-c30c-4c08-ab9e-ee202a858cb6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 992.223680] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aabb2335-d2ae-4788-b189-2d831bcd10cc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.230580] env[62753]: DEBUG nova.compute.manager [None req-3556fdab-c30c-4c08-ab9e-ee202a858cb6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62753) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 992.231202] env[62753]: DEBUG nova.objects.instance [None req-3556fdab-c30c-4c08-ab9e-ee202a858cb6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lazy-loading 'flavor' on Instance uuid 91fc519e-c9bb-4605-bdc6-4f463561b988 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 992.305840] env[62753]: DEBUG nova.compute.utils [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 992.310747] env[62753]: DEBUG nova.compute.manager [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 992.310978] env[62753]: DEBUG nova.network.neutron [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 992.355040] env[62753]: DEBUG nova.policy [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4dbc63095b0464fa0d2de86e3cf170e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5636da907ab343f9b42dbbd903d32283', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 992.510961] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d02d31b-3bd7-4920-b53d-34a9bb2f4abc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.518905] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae2ae37c-de7e-4db7-a132-5c53501cb5b5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.555791] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81c3cd52-43e0-4f81-add7-b76583dc85a9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.562765] env[62753]: DEBUG oslo_vmware.api [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332587, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.568575] env[62753]: DEBUG oslo_vmware.api [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332584, 'name': PowerOnVM_Task, 'duration_secs': 0.593806} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.570861] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 992.571088] env[62753]: INFO nova.compute.manager [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Took 7.52 seconds to spawn the instance on the hypervisor. [ 992.571293] env[62753]: DEBUG nova.compute.manager [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 992.572086] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35fa2272-4f53-4e63-b2c5-6ff546543f41 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.575600] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bab0795b-ece0-420d-8168-53c121b22adb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.582432] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332585, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.597845] env[62753]: DEBUG nova.compute.provider_tree [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 992.636399] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332586, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.555389} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 992.636399] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] e1c59139-ff02-453e-b54f-a56b1e40fdb5/e1c59139-ff02-453e-b54f-a56b1e40fdb5.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 992.636627] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 992.637254] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e3ee1c8f-707a-4c97-bf20-43e322bfcda5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.644900] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 992.644900] env[62753]: value = "task-1332588" [ 992.644900] env[62753]: _type = "Task" [ 992.644900] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.663089] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332588, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.738220] env[62753]: DEBUG nova.network.neutron [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Successfully created port: f9490496-ad16-4da2-805c-a5187d087d67 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 992.739445] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-3556fdab-c30c-4c08-ab9e-ee202a858cb6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 992.740532] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b0257bba-5a9c-493a-859d-4b942c82019e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 992.747475] env[62753]: DEBUG oslo_vmware.api [None req-3556fdab-c30c-4c08-ab9e-ee202a858cb6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 992.747475] env[62753]: value = "task-1332589" [ 992.747475] env[62753]: _type = "Task" [ 992.747475] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 992.756997] env[62753]: DEBUG oslo_vmware.api [None req-3556fdab-c30c-4c08-ab9e-ee202a858cb6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332589, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 992.812053] env[62753]: DEBUG nova.compute.manager [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 992.865484] env[62753]: DEBUG nova.network.neutron [req-3ccdbf2a-b459-4d12-aa26-f0d06167cb44 req-7a622bd5-f6d8-4d75-9cca-741035f172a7 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Updated VIF entry in instance network info cache for port 2498c430-c268-4e1e-b331-0a3b05eeae0a. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 992.865969] env[62753]: DEBUG nova.network.neutron [req-3ccdbf2a-b459-4d12-aa26-f0d06167cb44 req-7a622bd5-f6d8-4d75-9cca-741035f172a7 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Updating instance_info_cache with network_info: [{"id": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "address": "fa:16:3e:57:5d:75", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12ca481a-0c", "ovs_interfaceid": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2498c430-c268-4e1e-b331-0a3b05eeae0a", "address": "fa:16:3e:f3:a1:e8", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2498c430-c2", "ovs_interfaceid": "2498c430-c268-4e1e-b331-0a3b05eeae0a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 993.034668] env[62753]: DEBUG oslo_vmware.api [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332587, 'name': ReconfigVM_Task, 'duration_secs': 0.854785} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.035287] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.035542] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Reconfigured VM to attach interface {{(pid=62753) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 993.072309] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332585, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.105322] env[62753]: DEBUG nova.scheduler.client.report [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 993.112831] env[62753]: INFO nova.compute.manager [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Took 14.28 seconds to build instance. [ 993.156922] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332588, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.116947} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.157659] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 993.158435] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b163600-4045-4b1a-8025-675ce850d5b1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.184222] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Reconfiguring VM instance instance-00000059 to attach disk [datastore2] e1c59139-ff02-453e-b54f-a56b1e40fdb5/e1c59139-ff02-453e-b54f-a56b1e40fdb5.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 993.184550] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-7020415b-faf1-4f19-a631-42280c41b102 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.203288] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 993.203288] env[62753]: value = "task-1332590" [ 993.203288] env[62753]: _type = "Task" [ 993.203288] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 993.211983] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332590, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.256970] env[62753]: DEBUG oslo_vmware.api [None req-3556fdab-c30c-4c08-ab9e-ee202a858cb6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332589, 'name': PowerOffVM_Task, 'duration_secs': 0.463795} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.257278] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-3556fdab-c30c-4c08-ab9e-ee202a858cb6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 993.257463] env[62753]: DEBUG nova.compute.manager [None req-3556fdab-c30c-4c08-ab9e-ee202a858cb6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 993.258245] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13db8940-e76c-43f3-8c3b-0559c1d550ed {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.368717] env[62753]: DEBUG oslo_concurrency.lockutils [req-3ccdbf2a-b459-4d12-aa26-f0d06167cb44 req-7a622bd5-f6d8-4d75-9cca-741035f172a7 service nova] Releasing lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 993.541326] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2ff952b0-1cdd-4df5-8eb6-d3a0777d78fe tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "interface-c3679c66-2763-4a04-8d14-e103bd952798-2498c430-c268-4e1e-b331-0a3b05eeae0a" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.150s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.572644] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332585, 'name': CloneVM_Task, 'duration_secs': 1.858013} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 993.572951] env[62753]: INFO nova.virt.vmwareapi.vmops [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Created linked-clone VM from snapshot [ 993.573684] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4d9b4922-2b88-474d-9482-92444380e72e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.581381] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Uploading image 819354f4-0c00-4b42-b73f-45d4df8fe9dd {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 993.608172] env[62753]: DEBUG oslo_vmware.rw_handles [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 993.608172] env[62753]: value = "vm-284657" [ 993.608172] env[62753]: _type = "VirtualMachine" [ 993.608172] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 993.608592] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-bafbd9b9-d1cb-4738-a985-c01dd2d69341 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.610553] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.809s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.612568] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 6.516s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 993.618025] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e757f3ff-893f-4f9b-b11a-0fb07ebc796b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "04fd0022-f20a-4217-9e47-5381635b17a6" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.792s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.620417] env[62753]: DEBUG oslo_vmware.rw_handles [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lease: (returnval){ [ 993.620417] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a22d34-88c2-fccd-4ebc-8a7801b1f9a9" [ 993.620417] env[62753]: _type = "HttpNfcLease" [ 993.620417] env[62753]: } obtained for exporting VM: (result){ [ 993.620417] env[62753]: value = "vm-284657" [ 993.620417] env[62753]: _type = "VirtualMachine" [ 993.620417] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 993.620719] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the lease: (returnval){ [ 993.620719] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a22d34-88c2-fccd-4ebc-8a7801b1f9a9" [ 993.620719] env[62753]: _type = "HttpNfcLease" [ 993.620719] env[62753]: } to be ready. {{(pid=62753) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 993.629503] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 993.629503] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a22d34-88c2-fccd-4ebc-8a7801b1f9a9" [ 993.629503] env[62753]: _type = "HttpNfcLease" [ 993.629503] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 993.639419] env[62753]: INFO nova.scheduler.client.report [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Deleted allocations for instance 77cdd901-cc96-4a6f-a696-a65f54a96b1a [ 993.714432] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332590, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 993.769522] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3556fdab-c30c-4c08-ab9e-ee202a858cb6 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "91fc519e-c9bb-4605-bdc6-4f463561b988" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.548s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 993.821110] env[62753]: DEBUG nova.compute.manager [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 993.844555] env[62753]: DEBUG nova.virt.hardware [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 993.845041] env[62753]: DEBUG nova.virt.hardware [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 993.845320] env[62753]: DEBUG nova.virt.hardware [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 993.845631] env[62753]: DEBUG nova.virt.hardware [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 993.845918] env[62753]: DEBUG nova.virt.hardware [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 993.846204] env[62753]: DEBUG nova.virt.hardware [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 993.846549] env[62753]: DEBUG nova.virt.hardware [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 993.846811] env[62753]: DEBUG nova.virt.hardware [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 993.847216] env[62753]: DEBUG nova.virt.hardware [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 993.847495] env[62753]: DEBUG nova.virt.hardware [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 993.847789] env[62753]: DEBUG nova.virt.hardware [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 993.849409] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcdd3adc-fd83-4784-be8d-a6762ac2bb4e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 993.860437] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bec79a31-e628-427f-a9e8-ef89403373b0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.118935] env[62753]: INFO nova.compute.claims [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 994.142086] env[62753]: DEBUG nova.compute.manager [req-4e3004b5-63d0-4bd3-9627-3652118c445e req-5a3366c4-7778-4bc5-b58d-bf70db9b5c0f service nova] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Received event network-changed-011eae16-b648-4e2e-abfb-e424e05e41c2 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 994.142297] env[62753]: DEBUG nova.compute.manager [req-4e3004b5-63d0-4bd3-9627-3652118c445e req-5a3366c4-7778-4bc5-b58d-bf70db9b5c0f service nova] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Refreshing instance network info cache due to event network-changed-011eae16-b648-4e2e-abfb-e424e05e41c2. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 994.142935] env[62753]: DEBUG oslo_concurrency.lockutils [req-4e3004b5-63d0-4bd3-9627-3652118c445e req-5a3366c4-7778-4bc5-b58d-bf70db9b5c0f service nova] Acquiring lock "refresh_cache-04fd0022-f20a-4217-9e47-5381635b17a6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.142935] env[62753]: DEBUG oslo_concurrency.lockutils [req-4e3004b5-63d0-4bd3-9627-3652118c445e req-5a3366c4-7778-4bc5-b58d-bf70db9b5c0f service nova] Acquired lock "refresh_cache-04fd0022-f20a-4217-9e47-5381635b17a6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.142935] env[62753]: DEBUG nova.network.neutron [req-4e3004b5-63d0-4bd3-9627-3652118c445e req-5a3366c4-7778-4bc5-b58d-bf70db9b5c0f service nova] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Refreshing network info cache for port 011eae16-b648-4e2e-abfb-e424e05e41c2 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 994.144399] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 994.144399] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a22d34-88c2-fccd-4ebc-8a7801b1f9a9" [ 994.144399] env[62753]: _type = "HttpNfcLease" [ 994.144399] env[62753]: } is ready. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 994.144881] env[62753]: DEBUG oslo_vmware.rw_handles [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 994.144881] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a22d34-88c2-fccd-4ebc-8a7801b1f9a9" [ 994.144881] env[62753]: _type = "HttpNfcLease" [ 994.144881] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 994.145710] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb25728d-e4ac-42bc-b36f-10d27fab2bc1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.151980] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bfd7b405-1b67-422c-b5c0-dcc621dd030c tempest-AttachVolumeTestJSON-1902721549 tempest-AttachVolumeTestJSON-1902721549-project-member] Lock "77cdd901-cc96-4a6f-a696-a65f54a96b1a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.231s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.159275] env[62753]: DEBUG oslo_vmware.rw_handles [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5222029c-6ab8-e855-afd7-8d0eff8164dc/disk-0.vmdk from lease info. {{(pid=62753) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 994.159388] env[62753]: DEBUG oslo_vmware.rw_handles [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5222029c-6ab8-e855-afd7-8d0eff8164dc/disk-0.vmdk for reading. {{(pid=62753) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 994.236102] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332590, 'name': ReconfigVM_Task, 'duration_secs': 0.900682} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.236405] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Reconfigured VM instance instance-00000059 to attach disk [datastore2] e1c59139-ff02-453e-b54f-a56b1e40fdb5/e1c59139-ff02-453e-b54f-a56b1e40fdb5.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 994.237045] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1f23efcf-703d-4b89-964a-c994c57f7b10 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.244174] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 994.244174] env[62753]: value = "task-1332592" [ 994.244174] env[62753]: _type = "Task" [ 994.244174] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.252651] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332592, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.262905] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-9b7d64a9-011d-4d58-b917-b861580da7d6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.302707] env[62753]: DEBUG nova.compute.manager [req-a85ec67c-b6f6-426b-b759-be42e26207a0 req-cf95518c-5fbb-47ff-b6dc-dca8922de39f service nova] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Received event network-vif-plugged-f9490496-ad16-4da2-805c-a5187d087d67 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 994.302932] env[62753]: DEBUG oslo_concurrency.lockutils [req-a85ec67c-b6f6-426b-b759-be42e26207a0 req-cf95518c-5fbb-47ff-b6dc-dca8922de39f service nova] Acquiring lock "3009b1e4-53b1-4b12-8aaf-3c36b37c5d99-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.303171] env[62753]: DEBUG oslo_concurrency.lockutils [req-a85ec67c-b6f6-426b-b759-be42e26207a0 req-cf95518c-5fbb-47ff-b6dc-dca8922de39f service nova] Lock "3009b1e4-53b1-4b12-8aaf-3c36b37c5d99-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.303344] env[62753]: DEBUG oslo_concurrency.lockutils [req-a85ec67c-b6f6-426b-b759-be42e26207a0 req-cf95518c-5fbb-47ff-b6dc-dca8922de39f service nova] Lock "3009b1e4-53b1-4b12-8aaf-3c36b37c5d99-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.303521] env[62753]: DEBUG nova.compute.manager [req-a85ec67c-b6f6-426b-b759-be42e26207a0 req-cf95518c-5fbb-47ff-b6dc-dca8922de39f service nova] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] No waiting events found dispatching network-vif-plugged-f9490496-ad16-4da2-805c-a5187d087d67 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 994.303682] env[62753]: WARNING nova.compute.manager [req-a85ec67c-b6f6-426b-b759-be42e26207a0 req-cf95518c-5fbb-47ff-b6dc-dca8922de39f service nova] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Received unexpected event network-vif-plugged-f9490496-ad16-4da2-805c-a5187d087d67 for instance with vm_state building and task_state spawning. [ 994.353125] env[62753]: DEBUG nova.network.neutron [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Successfully updated port: f9490496-ad16-4da2-805c-a5187d087d67 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 994.628665] env[62753]: INFO nova.compute.resource_tracker [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Updating resource usage from migration 4a936932-16ef-44c3-898d-da10d29f220b [ 994.765508] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332592, 'name': Rename_Task, 'duration_secs': 0.188414} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 994.766758] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 994.766758] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4acddf8a-caa1-4c5e-a4ce-e6c9c48f1512 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.774194] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 994.774194] env[62753]: value = "task-1332594" [ 994.774194] env[62753]: _type = "Task" [ 994.774194] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 994.787070] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332594, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 994.854961] env[62753]: DEBUG oslo_concurrency.lockutils [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "refresh_cache-3009b1e4-53b1-4b12-8aaf-3c36b37c5d99" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 994.855144] env[62753]: DEBUG oslo_concurrency.lockutils [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "refresh_cache-3009b1e4-53b1-4b12-8aaf-3c36b37c5d99" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 994.855816] env[62753]: DEBUG nova.network.neutron [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 994.895616] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a4447070-346c-4f7a-b3bd-bf6befdef9ff {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.904780] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e0315a0-0fbb-4c20-87b4-6a366e1dbc18 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.939191] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "91fc519e-c9bb-4605-bdc6-4f463561b988" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.939610] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "91fc519e-c9bb-4605-bdc6-4f463561b988" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.939909] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "91fc519e-c9bb-4605-bdc6-4f463561b988-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 994.940159] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "91fc519e-c9bb-4605-bdc6-4f463561b988-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 994.940380] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "91fc519e-c9bb-4605-bdc6-4f463561b988-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 994.945672] env[62753]: INFO nova.compute.manager [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Terminating instance [ 994.947919] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-253ebcde-a348-4094-a0d9-6e52947d297b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.951063] env[62753]: DEBUG nova.compute.manager [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 994.951350] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 994.952509] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d67a89ae-6269-4318-84f8-c7de91594c13 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.964885] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-264bdb28-1425-4636-8277-5e956fa0a3e8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.967936] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 994.968590] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-63b9aa9a-d955-436f-81a6-3e2329af5dd9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 994.980086] env[62753]: DEBUG nova.compute.provider_tree [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 994.989066] env[62753]: DEBUG nova.network.neutron [req-4e3004b5-63d0-4bd3-9627-3652118c445e req-5a3366c4-7778-4bc5-b58d-bf70db9b5c0f service nova] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Updated VIF entry in instance network info cache for port 011eae16-b648-4e2e-abfb-e424e05e41c2. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 994.989480] env[62753]: DEBUG nova.network.neutron [req-4e3004b5-63d0-4bd3-9627-3652118c445e req-5a3366c4-7778-4bc5-b58d-bf70db9b5c0f service nova] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Updating instance_info_cache with network_info: [{"id": "011eae16-b648-4e2e-abfb-e424e05e41c2", "address": "fa:16:3e:32:28:cb", "network": {"id": "61c6e640-d2e6-4b3b-b191-7da9249dd339", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-703836479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.214", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8925b84dcf9a47fbaf2eb8044b3850fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap011eae16-b6", "ovs_interfaceid": "011eae16-b648-4e2e-abfb-e424e05e41c2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.046601] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 995.046972] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 995.047309] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Deleting the datastore file [datastore2] 91fc519e-c9bb-4605-bdc6-4f463561b988 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 995.047603] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-54b2da73-7a06-487f-bece-0f68e66dd954 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.055010] env[62753]: DEBUG oslo_vmware.api [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 995.055010] env[62753]: value = "task-1332596" [ 995.055010] env[62753]: _type = "Task" [ 995.055010] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.064582] env[62753]: DEBUG oslo_vmware.api [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332596, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.217812] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "interface-c3679c66-2763-4a04-8d14-e103bd952798-2498c430-c268-4e1e-b331-0a3b05eeae0a" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 995.218144] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "interface-c3679c66-2763-4a04-8d14-e103bd952798-2498c430-c268-4e1e-b331-0a3b05eeae0a" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 995.288268] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332594, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.392707] env[62753]: DEBUG nova.network.neutron [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 995.483892] env[62753]: DEBUG nova.scheduler.client.report [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 995.491617] env[62753]: DEBUG oslo_concurrency.lockutils [req-4e3004b5-63d0-4bd3-9627-3652118c445e req-5a3366c4-7778-4bc5-b58d-bf70db9b5c0f service nova] Releasing lock "refresh_cache-04fd0022-f20a-4217-9e47-5381635b17a6" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 995.535051] env[62753]: DEBUG nova.network.neutron [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Updating instance_info_cache with network_info: [{"id": "f9490496-ad16-4da2-805c-a5187d087d67", "address": "fa:16:3e:f3:38:04", "network": {"id": "35ed0ec0-821f-45b6-bc2d-015626bfc0a2", "bridge": "br-int", "label": "tempest-ServersTestJSON-444726172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5636da907ab343f9b42dbbd903d32283", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9490496-ad", "ovs_interfaceid": "f9490496-ad16-4da2-805c-a5187d087d67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 995.566251] env[62753]: DEBUG oslo_vmware.api [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332596, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.203362} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.566533] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 995.566967] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 995.567177] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 995.567459] env[62753]: INFO nova.compute.manager [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Took 0.62 seconds to destroy the instance on the hypervisor. [ 995.567814] env[62753]: DEBUG oslo.service.loopingcall [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 995.568084] env[62753]: DEBUG nova.compute.manager [-] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 995.568235] env[62753]: DEBUG nova.network.neutron [-] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 995.720479] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 995.720675] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 995.721856] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12794a21-0f26-4949-b366-fd5d5196e4be {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.743346] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40dfb358-0bec-4b4a-99ee-4b4e700d6257 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.775286] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Reconfiguring VM to detach interface {{(pid=62753) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 995.776160] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c8b1f9b4-8500-4771-9c1d-8f7a9f6f11a6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.800427] env[62753]: DEBUG oslo_vmware.api [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332594, 'name': PowerOnVM_Task, 'duration_secs': 0.748041} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 995.802047] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 995.802420] env[62753]: INFO nova.compute.manager [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Took 8.39 seconds to spawn the instance on the hypervisor. [ 995.802721] env[62753]: DEBUG nova.compute.manager [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 995.803182] env[62753]: DEBUG oslo_vmware.api [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 995.803182] env[62753]: value = "task-1332597" [ 995.803182] env[62753]: _type = "Task" [ 995.803182] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 995.804079] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c38c532-17cf-4769-b866-f6891a57e62b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 995.820318] env[62753]: DEBUG oslo_vmware.api [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332597, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 995.989696] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.377s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 995.989914] env[62753]: INFO nova.compute.manager [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Migrating [ 995.997952] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.267s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 996.000480] env[62753]: INFO nova.compute.claims [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 996.038037] env[62753]: DEBUG oslo_concurrency.lockutils [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "refresh_cache-3009b1e4-53b1-4b12-8aaf-3c36b37c5d99" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 996.038941] env[62753]: DEBUG nova.compute.manager [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Instance network_info: |[{"id": "f9490496-ad16-4da2-805c-a5187d087d67", "address": "fa:16:3e:f3:38:04", "network": {"id": "35ed0ec0-821f-45b6-bc2d-015626bfc0a2", "bridge": "br-int", "label": "tempest-ServersTestJSON-444726172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5636da907ab343f9b42dbbd903d32283", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9490496-ad", "ovs_interfaceid": "f9490496-ad16-4da2-805c-a5187d087d67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 996.039736] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f3:38:04', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2624812a-9f9c-461d-8b5f-79bea90c7ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f9490496-ad16-4da2-805c-a5187d087d67', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 996.051291] env[62753]: DEBUG oslo.service.loopingcall [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 996.052012] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 996.052388] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0a169c2c-5231-47cb-8fb4-3480f2de2710 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 996.085365] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 996.085365] env[62753]: value = "task-1332598" [ 996.085365] env[62753]: _type = "Task" [ 996.085365] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 996.094574] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332598, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.316852] env[62753]: DEBUG oslo_vmware.api [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332597, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.329723] env[62753]: INFO nova.compute.manager [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Took 16.26 seconds to build instance. [ 996.337268] env[62753]: DEBUG nova.compute.manager [req-7e0e4727-803a-4a74-9e57-9756204aff66 req-d0415dbb-8258-42de-8fdd-c38998966f3a service nova] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Received event network-changed-f9490496-ad16-4da2-805c-a5187d087d67 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 996.337268] env[62753]: DEBUG nova.compute.manager [req-7e0e4727-803a-4a74-9e57-9756204aff66 req-d0415dbb-8258-42de-8fdd-c38998966f3a service nova] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Refreshing instance network info cache due to event network-changed-f9490496-ad16-4da2-805c-a5187d087d67. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 996.337268] env[62753]: DEBUG oslo_concurrency.lockutils [req-7e0e4727-803a-4a74-9e57-9756204aff66 req-d0415dbb-8258-42de-8fdd-c38998966f3a service nova] Acquiring lock "refresh_cache-3009b1e4-53b1-4b12-8aaf-3c36b37c5d99" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.337437] env[62753]: DEBUG oslo_concurrency.lockutils [req-7e0e4727-803a-4a74-9e57-9756204aff66 req-d0415dbb-8258-42de-8fdd-c38998966f3a service nova] Acquired lock "refresh_cache-3009b1e4-53b1-4b12-8aaf-3c36b37c5d99" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.337616] env[62753]: DEBUG nova.network.neutron [req-7e0e4727-803a-4a74-9e57-9756204aff66 req-d0415dbb-8258-42de-8fdd-c38998966f3a service nova] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Refreshing network info cache for port f9490496-ad16-4da2-805c-a5187d087d67 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 996.521305] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "refresh_cache-2b90824a-5d49-4ec9-bcf6-4f868ca876cf" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 996.521492] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "refresh_cache-2b90824a-5d49-4ec9-bcf6-4f868ca876cf" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 996.521668] env[62753]: DEBUG nova.network.neutron [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 996.583378] env[62753]: DEBUG nova.network.neutron [-] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 996.597725] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332598, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.818794] env[62753]: DEBUG oslo_vmware.api [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332597, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 996.831512] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0a2f9474-3ec4-483e-9c71-521c6fe2cb2c tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "e1c59139-ff02-453e-b54f-a56b1e40fdb5" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.773s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 997.091211] env[62753]: INFO nova.compute.manager [-] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Took 1.52 seconds to deallocate network for instance. [ 997.107273] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332598, 'name': CreateVM_Task, 'duration_secs': 0.626284} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.107273] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 997.108521] env[62753]: DEBUG oslo_concurrency.lockutils [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.108704] env[62753]: DEBUG oslo_concurrency.lockutils [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.109370] env[62753]: DEBUG oslo_concurrency.lockutils [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 997.109867] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-78a6e529-ef62-43ee-bc99-195129a5cd3d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.115834] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 997.115834] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]524b0378-2f5b-e4ad-6c66-dc593b9f9aba" [ 997.115834] env[62753]: _type = "Task" [ 997.115834] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.125892] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]524b0378-2f5b-e4ad-6c66-dc593b9f9aba, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.130595] env[62753]: DEBUG nova.network.neutron [req-7e0e4727-803a-4a74-9e57-9756204aff66 req-d0415dbb-8258-42de-8fdd-c38998966f3a service nova] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Updated VIF entry in instance network info cache for port f9490496-ad16-4da2-805c-a5187d087d67. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 997.130978] env[62753]: DEBUG nova.network.neutron [req-7e0e4727-803a-4a74-9e57-9756204aff66 req-d0415dbb-8258-42de-8fdd-c38998966f3a service nova] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Updating instance_info_cache with network_info: [{"id": "f9490496-ad16-4da2-805c-a5187d087d67", "address": "fa:16:3e:f3:38:04", "network": {"id": "35ed0ec0-821f-45b6-bc2d-015626bfc0a2", "bridge": "br-int", "label": "tempest-ServersTestJSON-444726172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5636da907ab343f9b42dbbd903d32283", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf9490496-ad", "ovs_interfaceid": "f9490496-ad16-4da2-805c-a5187d087d67", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.250837] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf6d87da-28d0-426a-b30f-40016286fcd6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.259994] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7d986f6-4a7f-4dac-b569-77bd3cb10095 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.297262] env[62753]: DEBUG nova.network.neutron [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Updating instance_info_cache with network_info: [{"id": "c86749ca-c0b7-4c20-a32a-fed94bb47e4d", "address": "fa:16:3e:8e:d2:8a", "network": {"id": "2823647f-d32d-4a1c-9e02-7577016a260e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1594498922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6c8a6bc946a469fa85cc3ab80d7333a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc86749ca-c0", "ovs_interfaceid": "c86749ca-c0b7-4c20-a32a-fed94bb47e4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.299118] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5503dd82-fab5-4781-91b6-734214e2f2d3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.314472] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5361cd1f-87df-4cc3-8005-0c4775fa0bfc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.326613] env[62753]: DEBUG oslo_vmware.api [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332597, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.336550] env[62753]: DEBUG nova.compute.provider_tree [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 997.606747] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 997.632297] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]524b0378-2f5b-e4ad-6c66-dc593b9f9aba, 'name': SearchDatastore_Task, 'duration_secs': 0.020558} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 997.632749] env[62753]: DEBUG oslo_concurrency.lockutils [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.633141] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 997.633523] env[62753]: DEBUG oslo_concurrency.lockutils [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 997.633784] env[62753]: DEBUG oslo_concurrency.lockutils [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 997.634152] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 997.634927] env[62753]: DEBUG oslo_concurrency.lockutils [req-7e0e4727-803a-4a74-9e57-9756204aff66 req-d0415dbb-8258-42de-8fdd-c38998966f3a service nova] Releasing lock "refresh_cache-3009b1e4-53b1-4b12-8aaf-3c36b37c5d99" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.635313] env[62753]: DEBUG nova.compute.manager [req-7e0e4727-803a-4a74-9e57-9756204aff66 req-d0415dbb-8258-42de-8fdd-c38998966f3a service nova] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Received event network-vif-deleted-59ca12e1-f9c6-4e17-a496-248b15733019 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 997.635614] env[62753]: INFO nova.compute.manager [req-7e0e4727-803a-4a74-9e57-9756204aff66 req-d0415dbb-8258-42de-8fdd-c38998966f3a service nova] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Neutron deleted interface 59ca12e1-f9c6-4e17-a496-248b15733019; detaching it from the instance and deleting it from the info cache [ 997.635908] env[62753]: DEBUG nova.network.neutron [req-7e0e4727-803a-4a74-9e57-9756204aff66 req-d0415dbb-8258-42de-8fdd-c38998966f3a service nova] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 997.637460] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0f970a2b-2366-49d5-ba2a-c266fcdd7db5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.648903] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 997.649129] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 997.650139] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3557a9a6-c680-4490-9ea8-76e2b1709496 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 997.656518] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 997.656518] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e6d24f-99a2-32de-32b9-cb5ea6c71b59" [ 997.656518] env[62753]: _type = "Task" [ 997.656518] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 997.666016] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e6d24f-99a2-32de-32b9-cb5ea6c71b59, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.804609] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "refresh_cache-2b90824a-5d49-4ec9-bcf6-4f868ca876cf" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 997.819017] env[62753]: DEBUG oslo_vmware.api [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332597, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 997.840309] env[62753]: DEBUG nova.scheduler.client.report [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 998.142447] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-d5dcfc9f-f34f-4c27-90f3-8d18c02570bd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.153463] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-45cb14db-8a9a-4c1d-b8fc-72ef0f36f9d8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.172351] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e6d24f-99a2-32de-32b9-cb5ea6c71b59, 'name': SearchDatastore_Task, 'duration_secs': 0.035367} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.173190] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6c1c40a7-de54-4521-aa6e-058f45bffd19 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.185040] env[62753]: DEBUG nova.compute.manager [req-7e0e4727-803a-4a74-9e57-9756204aff66 req-d0415dbb-8258-42de-8fdd-c38998966f3a service nova] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Detach interface failed, port_id=59ca12e1-f9c6-4e17-a496-248b15733019, reason: Instance 91fc519e-c9bb-4605-bdc6-4f463561b988 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 998.186644] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 998.186644] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52664156-0b0e-f655-c366-7c1e04f1ca32" [ 998.186644] env[62753]: _type = "Task" [ 998.186644] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.195056] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52664156-0b0e-f655-c366-7c1e04f1ca32, 'name': SearchDatastore_Task, 'duration_secs': 0.014302} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 998.195346] env[62753]: DEBUG oslo_concurrency.lockutils [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 998.195607] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99/3009b1e4-53b1-4b12-8aaf-3c36b37c5d99.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 998.195895] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f9eec8e1-4613-4bcd-afe5-caf32cd40976 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.202330] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 998.202330] env[62753]: value = "task-1332600" [ 998.202330] env[62753]: _type = "Task" [ 998.202330] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 998.209953] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332600, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.318853] env[62753]: DEBUG oslo_vmware.api [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332597, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.345593] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.348s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 998.346197] env[62753]: DEBUG nova.compute.manager [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 998.348836] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 7.290s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 998.349110] env[62753]: DEBUG nova.objects.instance [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lazy-loading 'pci_requests' on Instance uuid 4d20e083-2959-453a-8875-47955bc02613 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 998.712679] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332600, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.792298] env[62753]: DEBUG nova.compute.manager [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 998.793150] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf620dd-8d5d-45aa-8ec6-b7d9f7210b1d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 998.820089] env[62753]: DEBUG oslo_vmware.api [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332597, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 998.853075] env[62753]: DEBUG nova.compute.utils [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 998.856408] env[62753]: DEBUG nova.objects.instance [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lazy-loading 'numa_topology' on Instance uuid 4d20e083-2959-453a-8875-47955bc02613 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 998.857806] env[62753]: DEBUG nova.compute.manager [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 998.857995] env[62753]: DEBUG nova.network.neutron [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 998.902824] env[62753]: DEBUG nova.policy [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '6996040b35454e028b94735be079b925', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '37368e15fa5a4a0a97ac1c22fcf0bb80', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 999.182956] env[62753]: DEBUG nova.network.neutron [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Successfully created port: 7f800ba5-2f01-46b7-b84a-88f900822bf2 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 999.213970] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332600, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.569101} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.214385] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99/3009b1e4-53b1-4b12-8aaf-3c36b37c5d99.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 999.214609] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 999.214890] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-798b971f-7703-464f-aef4-3e2a60549889 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.221648] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 999.221648] env[62753]: value = "task-1332601" [ 999.221648] env[62753]: _type = "Task" [ 999.221648] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.229539] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332601, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.305182] env[62753]: INFO nova.compute.manager [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] instance snapshotting [ 999.308385] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1bbdf45-0c39-4bc4-973a-5d05a452d2d1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.323174] env[62753]: DEBUG oslo_vmware.api [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332597, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.340245] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dc21ec0-2176-4d73-a021-6ef50eaecade {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.344051] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9917b7e-47d3-4d37-8fa4-021336c8cc50 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.366140] env[62753]: DEBUG nova.compute.manager [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 999.372243] env[62753]: INFO nova.compute.claims [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 999.378586] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Updating instance '2b90824a-5d49-4ec9-bcf6-4f868ca876cf' progress to 0 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 999.731989] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332601, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.083172} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 999.732319] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 999.733123] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ad224598-4cc3-4dc2-bd88-b47d0b1a53e0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.757471] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Reconfiguring VM instance instance-0000005a to attach disk [datastore2] 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99/3009b1e4-53b1-4b12-8aaf-3c36b37c5d99.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 999.757797] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-717177f6-4a60-44a6-a087-8d0afffe112d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.778499] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 999.778499] env[62753]: value = "task-1332602" [ 999.778499] env[62753]: _type = "Task" [ 999.778499] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.786558] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332602, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.821088] env[62753]: DEBUG oslo_vmware.api [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332597, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.883778] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Creating Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 999.889321] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 999.889989] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-123bd4b7-7a3b-4621-90d8-b0dade273db4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.893066] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-fc18cd90-5f80-4b64-9bad-00a11f515f5a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 999.900177] env[62753]: DEBUG oslo_vmware.api [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 999.900177] env[62753]: value = "task-1332604" [ 999.900177] env[62753]: _type = "Task" [ 999.900177] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.901697] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 999.901697] env[62753]: value = "task-1332603" [ 999.901697] env[62753]: _type = "Task" [ 999.901697] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 999.917755] env[62753]: DEBUG oslo_vmware.api [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332604, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 999.921383] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332603, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.290416] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332602, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.321971] env[62753]: DEBUG oslo_vmware.api [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332597, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.391676] env[62753]: DEBUG nova.compute.manager [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1000.417805] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332603, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.418128] env[62753]: DEBUG oslo_vmware.api [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332604, 'name': PowerOffVM_Task, 'duration_secs': 0.236218} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.422402] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1000.422605] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Updating instance '2b90824a-5d49-4ec9-bcf6-4f868ca876cf' progress to 17 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1000.435043] env[62753]: DEBUG nova.virt.hardware [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1000.435171] env[62753]: DEBUG nova.virt.hardware [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1000.435263] env[62753]: DEBUG nova.virt.hardware [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1000.435450] env[62753]: DEBUG nova.virt.hardware [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1000.435620] env[62753]: DEBUG nova.virt.hardware [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1000.435789] env[62753]: DEBUG nova.virt.hardware [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1000.436221] env[62753]: DEBUG nova.virt.hardware [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1000.436441] env[62753]: DEBUG nova.virt.hardware [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1000.436631] env[62753]: DEBUG nova.virt.hardware [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1000.437345] env[62753]: DEBUG nova.virt.hardware [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1000.437345] env[62753]: DEBUG nova.virt.hardware [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1000.437831] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff82478e-c23c-4423-97c2-2a13049bf86b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.450799] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c78e2861-c0ac-4566-91a4-cd6f73cb1739 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.610859] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cb48afbe-b1c5-4e7b-8d2d-0ccb200e4de5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.617629] env[62753]: DEBUG nova.compute.manager [req-cbae162c-d866-4bbe-9379-b39e40bfdcd1 req-9e8aa868-4ecf-45ca-a55b-88165d02077c service nova] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Received event network-vif-plugged-7f800ba5-2f01-46b7-b84a-88f900822bf2 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1000.617963] env[62753]: DEBUG oslo_concurrency.lockutils [req-cbae162c-d866-4bbe-9379-b39e40bfdcd1 req-9e8aa868-4ecf-45ca-a55b-88165d02077c service nova] Acquiring lock "9d7066e1-3721-42d2-8e80-91eacbbd8b80-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1000.618331] env[62753]: DEBUG oslo_concurrency.lockutils [req-cbae162c-d866-4bbe-9379-b39e40bfdcd1 req-9e8aa868-4ecf-45ca-a55b-88165d02077c service nova] Lock "9d7066e1-3721-42d2-8e80-91eacbbd8b80-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1000.618572] env[62753]: DEBUG oslo_concurrency.lockutils [req-cbae162c-d866-4bbe-9379-b39e40bfdcd1 req-9e8aa868-4ecf-45ca-a55b-88165d02077c service nova] Lock "9d7066e1-3721-42d2-8e80-91eacbbd8b80-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1000.619072] env[62753]: DEBUG nova.compute.manager [req-cbae162c-d866-4bbe-9379-b39e40bfdcd1 req-9e8aa868-4ecf-45ca-a55b-88165d02077c service nova] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] No waiting events found dispatching network-vif-plugged-7f800ba5-2f01-46b7-b84a-88f900822bf2 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1000.619296] env[62753]: WARNING nova.compute.manager [req-cbae162c-d866-4bbe-9379-b39e40bfdcd1 req-9e8aa868-4ecf-45ca-a55b-88165d02077c service nova] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Received unexpected event network-vif-plugged-7f800ba5-2f01-46b7-b84a-88f900822bf2 for instance with vm_state building and task_state spawning. [ 1000.624033] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-07d40af8-85d6-4048-a7b6-2a5b61e2d9d2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.662243] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b525975-51f2-47db-a474-389847b3ed48 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.672642] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b00c988b-9001-4687-bdd4-6e905e44ae23 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.687130] env[62753]: DEBUG nova.compute.provider_tree [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1000.729320] env[62753]: DEBUG nova.network.neutron [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Successfully updated port: 7f800ba5-2f01-46b7-b84a-88f900822bf2 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1000.789945] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332602, 'name': ReconfigVM_Task, 'duration_secs': 0.522474} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1000.790439] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Reconfigured VM instance instance-0000005a to attach disk [datastore2] 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99/3009b1e4-53b1-4b12-8aaf-3c36b37c5d99.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1000.791100] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-9bee4244-3223-4b47-b2ff-0e6617fa8f3e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.798496] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 1000.798496] env[62753]: value = "task-1332605" [ 1000.798496] env[62753]: _type = "Task" [ 1000.798496] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.807574] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332605, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.823588] env[62753]: DEBUG oslo_vmware.api [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332597, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.919036] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332603, 'name': CreateSnapshot_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1000.932390] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1000.932632] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1000.933155] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1000.933288] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1000.933487] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1000.933845] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1000.934049] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1000.934593] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1000.934593] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1000.934918] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1000.935605] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1000.941852] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6eba10fc-5c0b-4ad2-b03d-672a4d232fd4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1000.959957] env[62753]: DEBUG oslo_vmware.api [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 1000.959957] env[62753]: value = "task-1332606" [ 1000.959957] env[62753]: _type = "Task" [ 1000.959957] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1000.969720] env[62753]: DEBUG oslo_vmware.api [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332606, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.190038] env[62753]: DEBUG nova.scheduler.client.report [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1001.232013] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "refresh_cache-9d7066e1-3721-42d2-8e80-91eacbbd8b80" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1001.232181] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquired lock "refresh_cache-9d7066e1-3721-42d2-8e80-91eacbbd8b80" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1001.232327] env[62753]: DEBUG nova.network.neutron [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1001.309842] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332605, 'name': Rename_Task, 'duration_secs': 0.200556} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.310178] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1001.310440] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-01d35ff9-f164-4d3d-a69a-6db524735f3c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.319010] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 1001.319010] env[62753]: value = "task-1332607" [ 1001.319010] env[62753]: _type = "Task" [ 1001.319010] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.326379] env[62753]: DEBUG oslo_vmware.api [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332597, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.331118] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332607, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.416512] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332603, 'name': CreateSnapshot_Task, 'duration_secs': 1.065473} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.416786] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Created Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1001.417465] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e80c0357-81bb-4692-b5f3-60ff83988064 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.471742] env[62753]: DEBUG oslo_vmware.api [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332606, 'name': ReconfigVM_Task, 'duration_secs': 0.402077} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.472196] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Updating instance '2b90824a-5d49-4ec9-bcf6-4f868ca876cf' progress to 33 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1001.695532] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.347s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1001.697776] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.091s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1001.698032] env[62753]: DEBUG nova.objects.instance [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lazy-loading 'resources' on Instance uuid 91fc519e-c9bb-4605-bdc6-4f463561b988 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1001.739367] env[62753]: INFO nova.network.neutron [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Updating port 9c52b1e4-82d2-4449-91f3-a624596bdc79 with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1001.781252] env[62753]: DEBUG nova.network.neutron [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1001.832150] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332607, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.837484] env[62753]: DEBUG oslo_vmware.api [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332597, 'name': ReconfigVM_Task, 'duration_secs': 5.784601} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1001.837801] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1001.838027] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Reconfigured VM to detach interface {{(pid=62753) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1001.942137] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Creating linked-clone VM from snapshot {{(pid=62753) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1001.942619] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-de15dc3a-05fe-4087-b3b4-4704bbae322a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1001.956208] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 1001.956208] env[62753]: value = "task-1332608" [ 1001.956208] env[62753]: _type = "Task" [ 1001.956208] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1001.965385] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332608, 'name': CloneVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1001.982252] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1001.982554] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1001.982777] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1001.982991] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1001.983581] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1001.983828] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1001.984202] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1001.984827] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1001.985092] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1001.985319] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1001.985512] env[62753]: DEBUG nova.virt.hardware [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1001.994697] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Reconfiguring VM instance instance-00000056 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1002.000899] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-47f3517a-7ae4-4f23-a947-b6b5249de6f1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.026944] env[62753]: DEBUG oslo_vmware.api [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 1002.026944] env[62753]: value = "task-1332609" [ 1002.026944] env[62753]: _type = "Task" [ 1002.026944] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.037561] env[62753]: DEBUG oslo_vmware.api [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332609, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.140184] env[62753]: DEBUG nova.network.neutron [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Updating instance_info_cache with network_info: [{"id": "7f800ba5-2f01-46b7-b84a-88f900822bf2", "address": "fa:16:3e:03:90:c5", "network": {"id": "22a90636-182f-4944-893f-329f446b07c4", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1513076888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "37368e15fa5a4a0a97ac1c22fcf0bb80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f800ba5-2f", "ovs_interfaceid": "7f800ba5-2f01-46b7-b84a-88f900822bf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.331097] env[62753]: DEBUG oslo_vmware.api [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332607, 'name': PowerOnVM_Task, 'duration_secs': 0.773357} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.331097] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1002.331097] env[62753]: INFO nova.compute.manager [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Took 8.51 seconds to spawn the instance on the hypervisor. [ 1002.331097] env[62753]: DEBUG nova.compute.manager [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1002.334616] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed9d3d96-261c-4161-af1f-7bc9fd9d8152 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.424856] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41a7fc82-78ff-4eab-9766-28451f5d92e1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.433197] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-448f77d3-f7b7-4f03-8c6a-882511efc64f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.470760] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dfbed64-7897-4bf4-b753-a59886c5237f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.479375] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332608, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.482593] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c51dd93-deba-412d-ab1c-d30edfb9b525 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.497779] env[62753]: DEBUG nova.compute.provider_tree [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1002.537048] env[62753]: DEBUG oslo_vmware.api [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332609, 'name': ReconfigVM_Task, 'duration_secs': 0.226223} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1002.537368] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Reconfigured VM instance instance-00000056 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1002.538215] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0885f4ef-ba7d-4fb4-974e-adfa02871c45 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.570566] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Reconfiguring VM instance instance-00000056 to attach disk [datastore2] 2b90824a-5d49-4ec9-bcf6-4f868ca876cf/2b90824a-5d49-4ec9-bcf6-4f868ca876cf.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1002.571114] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-edd45da9-51d0-495c-8e28-53f3aad84ff3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.601138] env[62753]: DEBUG oslo_vmware.api [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 1002.601138] env[62753]: value = "task-1332610" [ 1002.601138] env[62753]: _type = "Task" [ 1002.601138] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.616788] env[62753]: DEBUG oslo_vmware.api [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332610, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.644649] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Releasing lock "refresh_cache-9d7066e1-3721-42d2-8e80-91eacbbd8b80" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1002.644649] env[62753]: DEBUG nova.compute.manager [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Instance network_info: |[{"id": "7f800ba5-2f01-46b7-b84a-88f900822bf2", "address": "fa:16:3e:03:90:c5", "network": {"id": "22a90636-182f-4944-893f-329f446b07c4", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1513076888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "37368e15fa5a4a0a97ac1c22fcf0bb80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f800ba5-2f", "ovs_interfaceid": "7f800ba5-2f01-46b7-b84a-88f900822bf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1002.645249] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:03:90:c5', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '8e7f6f41-f4eb-4832-a390-730fca1cf717', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7f800ba5-2f01-46b7-b84a-88f900822bf2', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1002.655191] env[62753]: DEBUG oslo.service.loopingcall [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1002.655191] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1002.656763] env[62753]: DEBUG nova.compute.manager [req-572197a6-a2ab-4733-b487-7582afa80b4c req-da7e466b-7b67-4e12-8cef-e56dc6bea127 service nova] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Received event network-changed-7f800ba5-2f01-46b7-b84a-88f900822bf2 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1002.656763] env[62753]: DEBUG nova.compute.manager [req-572197a6-a2ab-4733-b487-7582afa80b4c req-da7e466b-7b67-4e12-8cef-e56dc6bea127 service nova] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Refreshing instance network info cache due to event network-changed-7f800ba5-2f01-46b7-b84a-88f900822bf2. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1002.656763] env[62753]: DEBUG oslo_concurrency.lockutils [req-572197a6-a2ab-4733-b487-7582afa80b4c req-da7e466b-7b67-4e12-8cef-e56dc6bea127 service nova] Acquiring lock "refresh_cache-9d7066e1-3721-42d2-8e80-91eacbbd8b80" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1002.656763] env[62753]: DEBUG oslo_concurrency.lockutils [req-572197a6-a2ab-4733-b487-7582afa80b4c req-da7e466b-7b67-4e12-8cef-e56dc6bea127 service nova] Acquired lock "refresh_cache-9d7066e1-3721-42d2-8e80-91eacbbd8b80" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1002.656763] env[62753]: DEBUG nova.network.neutron [req-572197a6-a2ab-4733-b487-7582afa80b4c req-da7e466b-7b67-4e12-8cef-e56dc6bea127 service nova] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Refreshing network info cache for port 7f800ba5-2f01-46b7-b84a-88f900822bf2 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1002.657718] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-be16e4cb-cdc4-480b-b711-fc2cb9cb8949 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1002.683354] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1002.683354] env[62753]: value = "task-1332611" [ 1002.683354] env[62753]: _type = "Task" [ 1002.683354] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1002.696620] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332611, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1002.855032] env[62753]: INFO nova.compute.manager [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Took 19.15 seconds to build instance. [ 1002.933274] env[62753]: DEBUG nova.network.neutron [req-572197a6-a2ab-4733-b487-7582afa80b4c req-da7e466b-7b67-4e12-8cef-e56dc6bea127 service nova] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Updated VIF entry in instance network info cache for port 7f800ba5-2f01-46b7-b84a-88f900822bf2. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1002.933810] env[62753]: DEBUG nova.network.neutron [req-572197a6-a2ab-4733-b487-7582afa80b4c req-da7e466b-7b67-4e12-8cef-e56dc6bea127 service nova] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Updating instance_info_cache with network_info: [{"id": "7f800ba5-2f01-46b7-b84a-88f900822bf2", "address": "fa:16:3e:03:90:c5", "network": {"id": "22a90636-182f-4944-893f-329f446b07c4", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1513076888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "37368e15fa5a4a0a97ac1c22fcf0bb80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f800ba5-2f", "ovs_interfaceid": "7f800ba5-2f01-46b7-b84a-88f900822bf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1002.976900] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332608, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.001066] env[62753]: DEBUG nova.scheduler.client.report [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1003.111250] env[62753]: DEBUG oslo_vmware.api [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332610, 'name': ReconfigVM_Task, 'duration_secs': 0.392589} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.111516] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Reconfigured VM instance instance-00000056 to attach disk [datastore2] 2b90824a-5d49-4ec9-bcf6-4f868ca876cf/2b90824a-5d49-4ec9-bcf6-4f868ca876cf.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1003.111795] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Updating instance '2b90824a-5d49-4ec9-bcf6-4f868ca876cf' progress to 50 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1003.138373] env[62753]: DEBUG oslo_vmware.rw_handles [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5222029c-6ab8-e855-afd7-8d0eff8164dc/disk-0.vmdk. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1003.140309] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9a7f359-7deb-40e1-bb24-9b9bd3958a4d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.147065] env[62753]: DEBUG oslo_vmware.rw_handles [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5222029c-6ab8-e855-afd7-8d0eff8164dc/disk-0.vmdk is in state: ready. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1003.147315] env[62753]: ERROR oslo_vmware.rw_handles [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5222029c-6ab8-e855-afd7-8d0eff8164dc/disk-0.vmdk due to incomplete transfer. [ 1003.147946] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-51e93dcc-b906-4b67-85a1-9bf72fc28173 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.156744] env[62753]: DEBUG oslo_vmware.rw_handles [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/5222029c-6ab8-e855-afd7-8d0eff8164dc/disk-0.vmdk. {{(pid=62753) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1003.156744] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Uploaded image 819354f4-0c00-4b42-b73f-45d4df8fe9dd to the Glance image server {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1003.158977] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Destroying the VM {{(pid=62753) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1003.159484] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-6463e7c8-f459-491a-af0d-118ca66a3d60 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.165851] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1003.165851] env[62753]: value = "task-1332612" [ 1003.165851] env[62753]: _type = "Task" [ 1003.165851] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.173663] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332612, 'name': Destroy_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.193598] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332611, 'name': CreateVM_Task, 'duration_secs': 0.354608} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.193767] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1003.194517] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.194684] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.195068] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1003.195597] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-293b0cbe-cd2d-49ca-a8b1-caafdf5a8360 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.200232] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1003.200232] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c850f2-ec91-9e80-1933-b2a8cfc028c5" [ 1003.200232] env[62753]: _type = "Task" [ 1003.200232] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.208318] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c850f2-ec91-9e80-1933-b2a8cfc028c5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.357290] env[62753]: DEBUG oslo_concurrency.lockutils [None req-54a4c00e-603d-4c24-85f0-ae0c28571a93 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "3009b1e4-53b1-4b12-8aaf-3c36b37c5d99" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.662s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.436690] env[62753]: DEBUG oslo_concurrency.lockutils [req-572197a6-a2ab-4733-b487-7582afa80b4c req-da7e466b-7b67-4e12-8cef-e56dc6bea127 service nova] Releasing lock "refresh_cache-9d7066e1-3721-42d2-8e80-91eacbbd8b80" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.443031] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.443031] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.443031] env[62753]: DEBUG nova.network.neutron [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1003.476537] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332608, 'name': CloneVM_Task, 'duration_secs': 1.377792} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.477183] env[62753]: INFO nova.virt.vmwareapi.vmops [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Created linked-clone VM from snapshot [ 1003.477966] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a63b9cbe-0a9d-4da7-b4a7-a785406aa7b8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.486229] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Uploading image 701eee2c-daa2-4a72-a40a-6c24d5ec6909 {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1003.496515] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Destroying the VM {{(pid=62753) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1003.496771] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-bd72b03f-8c30-4277-b6cb-2c2959cad0d9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.503310] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 1003.503310] env[62753]: value = "task-1332613" [ 1003.503310] env[62753]: _type = "Task" [ 1003.503310] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.507549] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.810s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1003.518339] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332613, 'name': Destroy_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.529595] env[62753]: INFO nova.scheduler.client.report [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Deleted allocations for instance 91fc519e-c9bb-4605-bdc6-4f463561b988 [ 1003.581054] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.581270] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquired lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.581449] env[62753]: DEBUG nova.network.neutron [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1003.623259] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c450c876-e214-4157-a8e5-0069e7548a92 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.654250] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6d8e323-eb20-4813-b84f-2dc560ba0379 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.688536] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Updating instance '2b90824a-5d49-4ec9-bcf6-4f868ca876cf' progress to 67 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1003.704918] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332612, 'name': Destroy_Task, 'duration_secs': 0.46034} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.709733] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Destroyed the VM [ 1003.710312] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Deleting Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1003.711485] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-cba092af-8ee4-4124-b031-06d037c779f3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.721292] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c850f2-ec91-9e80-1933-b2a8cfc028c5, 'name': SearchDatastore_Task, 'duration_secs': 0.011355} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1003.723177] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1003.723519] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1003.723915] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1003.724187] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1003.724474] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1003.724986] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1003.724986] env[62753]: value = "task-1332614" [ 1003.724986] env[62753]: _type = "Task" [ 1003.724986] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.725276] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-35961732-5260-4f2f-ab4b-182c3a0ef306 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.740751] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332614, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1003.742168] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1003.742421] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1003.743465] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fa92c10d-e312-40c8-b560-08651a22387f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1003.748841] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1003.748841] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52060540-5002-b5f0-9016-f48c4556cd57" [ 1003.748841] env[62753]: _type = "Task" [ 1003.748841] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1003.757224] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52060540-5002-b5f0-9016-f48c4556cd57, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.010228] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1004.014325] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332613, 'name': Destroy_Task} progress is 33%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.042904] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5aea9de2-5967-4ee3-b63a-9e37cc7140af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "91fc519e-c9bb-4605-bdc6-4f463561b988" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.103s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.240360] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332614, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.258696] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52060540-5002-b5f0-9016-f48c4556cd57, 'name': SearchDatastore_Task, 'duration_secs': 0.038264} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.259692] env[62753]: INFO nova.network.neutron [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Port 2498c430-c268-4e1e-b331-0a3b05eeae0a from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1004.260181] env[62753]: DEBUG nova.network.neutron [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Updating instance_info_cache with network_info: [{"id": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "address": "fa:16:3e:57:5d:75", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12ca481a-0c", "ovs_interfaceid": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.262454] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9474c897-515b-47a1-b52c-022f4db9e7bc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.268810] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1004.268810] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]528d76d5-7d97-1086-bbd9-fb4ce12ac9b6" [ 1004.268810] env[62753]: _type = "Task" [ 1004.268810] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.278599] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]528d76d5-7d97-1086-bbd9-fb4ce12ac9b6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.292732] env[62753]: DEBUG nova.network.neutron [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Port c86749ca-c0b7-4c20-a32a-fed94bb47e4d binding to destination host cpu-1 is already ACTIVE {{(pid=62753) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1004.391384] env[62753]: DEBUG nova.network.neutron [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Updating instance_info_cache with network_info: [{"id": "9c52b1e4-82d2-4449-91f3-a624596bdc79", "address": "fa:16:3e:8e:1b:de", "network": {"id": "12446a5e-7329-4f8f-aa28-a8b9d78e7573", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1766107072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "371a5e8b16be4cffbdeb2139757684fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7c7f16b-a2e3-4d1f-9b7e-a44a9ebb589c", "external-id": "nsx-vlan-transportzone-772", "segmentation_id": 772, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c52b1e4-82", "ovs_interfaceid": "9c52b1e4-82d2-4449-91f3-a624596bdc79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1004.425547] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "3009b1e4-53b1-4b12-8aaf-3c36b37c5d99" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.425816] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "3009b1e4-53b1-4b12-8aaf-3c36b37c5d99" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.426035] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "3009b1e4-53b1-4b12-8aaf-3c36b37c5d99-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.426270] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "3009b1e4-53b1-4b12-8aaf-3c36b37c5d99-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.426461] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "3009b1e4-53b1-4b12-8aaf-3c36b37c5d99-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.428740] env[62753]: INFO nova.compute.manager [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Terminating instance [ 1004.431351] env[62753]: DEBUG nova.compute.manager [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1004.431568] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1004.432478] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6470a60-218d-4f71-985a-fddef4886036 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.439949] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1004.440266] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5f9f67a8-ddeb-46f6-b823-de12d5f89644 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.446427] env[62753]: DEBUG oslo_vmware.api [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 1004.446427] env[62753]: value = "task-1332615" [ 1004.446427] env[62753]: _type = "Task" [ 1004.446427] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.453947] env[62753]: DEBUG oslo_vmware.api [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332615, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.514328] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332613, 'name': Destroy_Task, 'duration_secs': 0.76345} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.514591] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Destroyed the VM [ 1004.514829] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Deleting Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1004.515137] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-363a8842-084c-4c4d-b8b9-62d965ba1303 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.521376] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 1004.521376] env[62753]: value = "task-1332616" [ 1004.521376] env[62753]: _type = "Task" [ 1004.521376] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.530215] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332616, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.688636] env[62753]: DEBUG nova.compute.manager [req-5740a8a1-ccfc-437f-a7a3-c518777cf675 req-ab3db403-0342-4c98-9d1b-f2296b7f4ff5 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Received event network-vif-plugged-9c52b1e4-82d2-4449-91f3-a624596bdc79 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.688836] env[62753]: DEBUG oslo_concurrency.lockutils [req-5740a8a1-ccfc-437f-a7a3-c518777cf675 req-ab3db403-0342-4c98-9d1b-f2296b7f4ff5 service nova] Acquiring lock "4d20e083-2959-453a-8875-47955bc02613-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1004.689061] env[62753]: DEBUG oslo_concurrency.lockutils [req-5740a8a1-ccfc-437f-a7a3-c518777cf675 req-ab3db403-0342-4c98-9d1b-f2296b7f4ff5 service nova] Lock "4d20e083-2959-453a-8875-47955bc02613-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1004.689241] env[62753]: DEBUG oslo_concurrency.lockutils [req-5740a8a1-ccfc-437f-a7a3-c518777cf675 req-ab3db403-0342-4c98-9d1b-f2296b7f4ff5 service nova] Lock "4d20e083-2959-453a-8875-47955bc02613-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1004.689416] env[62753]: DEBUG nova.compute.manager [req-5740a8a1-ccfc-437f-a7a3-c518777cf675 req-ab3db403-0342-4c98-9d1b-f2296b7f4ff5 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] No waiting events found dispatching network-vif-plugged-9c52b1e4-82d2-4449-91f3-a624596bdc79 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1004.689587] env[62753]: WARNING nova.compute.manager [req-5740a8a1-ccfc-437f-a7a3-c518777cf675 req-ab3db403-0342-4c98-9d1b-f2296b7f4ff5 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Received unexpected event network-vif-plugged-9c52b1e4-82d2-4449-91f3-a624596bdc79 for instance with vm_state shelved_offloaded and task_state spawning. [ 1004.689746] env[62753]: DEBUG nova.compute.manager [req-5740a8a1-ccfc-437f-a7a3-c518777cf675 req-ab3db403-0342-4c98-9d1b-f2296b7f4ff5 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Received event network-changed-9c52b1e4-82d2-4449-91f3-a624596bdc79 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.689917] env[62753]: DEBUG nova.compute.manager [req-5740a8a1-ccfc-437f-a7a3-c518777cf675 req-ab3db403-0342-4c98-9d1b-f2296b7f4ff5 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Refreshing instance network info cache due to event network-changed-9c52b1e4-82d2-4449-91f3-a624596bdc79. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1004.692873] env[62753]: DEBUG oslo_concurrency.lockutils [req-5740a8a1-ccfc-437f-a7a3-c518777cf675 req-ab3db403-0342-4c98-9d1b-f2296b7f4ff5 service nova] Acquiring lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.713608] env[62753]: DEBUG nova.compute.manager [req-61133bb0-f347-427a-83e0-eee0f5646543 req-ad2dc458-37a4-4803-b99d-6c5c14b6f6fe service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Received event network-changed-12ca481a-0cb3-465e-b04c-802ae3ce6438 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1004.714087] env[62753]: DEBUG nova.compute.manager [req-61133bb0-f347-427a-83e0-eee0f5646543 req-ad2dc458-37a4-4803-b99d-6c5c14b6f6fe service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Refreshing instance network info cache due to event network-changed-12ca481a-0cb3-465e-b04c-802ae3ce6438. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1004.714412] env[62753]: DEBUG oslo_concurrency.lockutils [req-61133bb0-f347-427a-83e0-eee0f5646543 req-ad2dc458-37a4-4803-b99d-6c5c14b6f6fe service nova] Acquiring lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1004.739658] env[62753]: DEBUG oslo_vmware.api [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332614, 'name': RemoveSnapshot_Task, 'duration_secs': 0.729136} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.739927] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Deleted Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1004.740196] env[62753]: INFO nova.compute.manager [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Took 15.30 seconds to snapshot the instance on the hypervisor. [ 1004.765825] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.767909] env[62753]: DEBUG oslo_concurrency.lockutils [req-61133bb0-f347-427a-83e0-eee0f5646543 req-ad2dc458-37a4-4803-b99d-6c5c14b6f6fe service nova] Acquired lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.768131] env[62753]: DEBUG nova.network.neutron [req-61133bb0-f347-427a-83e0-eee0f5646543 req-ad2dc458-37a4-4803-b99d-6c5c14b6f6fe service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Refreshing network info cache for port 12ca481a-0cb3-465e-b04c-802ae3ce6438 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1004.778944] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]528d76d5-7d97-1086-bbd9-fb4ce12ac9b6, 'name': SearchDatastore_Task, 'duration_secs': 0.010249} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.779296] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.779479] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 9d7066e1-3721-42d2-8e80-91eacbbd8b80/9d7066e1-3721-42d2-8e80-91eacbbd8b80.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1004.779970] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-df9497a3-52f4-4b0b-94f7-411f9bed437c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.789238] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1004.789238] env[62753]: value = "task-1332617" [ 1004.789238] env[62753]: _type = "Task" [ 1004.789238] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.802085] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332617, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1004.895174] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Releasing lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1004.899174] env[62753]: DEBUG oslo_concurrency.lockutils [req-5740a8a1-ccfc-437f-a7a3-c518777cf675 req-ab3db403-0342-4c98-9d1b-f2296b7f4ff5 service nova] Acquired lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1004.899344] env[62753]: DEBUG nova.network.neutron [req-5740a8a1-ccfc-437f-a7a3-c518777cf675 req-ab3db403-0342-4c98-9d1b-f2296b7f4ff5 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Refreshing network info cache for port 9c52b1e4-82d2-4449-91f3-a624596bdc79 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1004.922560] env[62753]: DEBUG nova.virt.hardware [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='6c4602124d4d38313a43bbe1c2668579',container_format='bare',created_at=2024-10-21T20:33:03Z,direct_url=,disk_format='vmdk',id=3b2f7019-a7ab-4e0e-9ea9-07c8a7334614,min_disk=1,min_ram=0,name='tempest-ServersNegativeTestJSON-server-17630419-shelved',owner='371a5e8b16be4cffbdeb2139757684fc',properties=ImageMetaProps,protected=,size=31664640,status='active',tags=,updated_at=2024-10-21T20:33:21Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1004.923024] env[62753]: DEBUG nova.virt.hardware [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1004.923024] env[62753]: DEBUG nova.virt.hardware [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1004.923165] env[62753]: DEBUG nova.virt.hardware [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1004.923301] env[62753]: DEBUG nova.virt.hardware [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1004.923450] env[62753]: DEBUG nova.virt.hardware [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1004.923665] env[62753]: DEBUG nova.virt.hardware [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1004.923827] env[62753]: DEBUG nova.virt.hardware [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1004.924048] env[62753]: DEBUG nova.virt.hardware [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1004.924234] env[62753]: DEBUG nova.virt.hardware [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1004.924410] env[62753]: DEBUG nova.virt.hardware [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1004.925544] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9604e1fa-b268-4478-b392-b827b5221673 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.934649] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac8a030-c212-40d3-89a9-b174d8fb02e3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.949464] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8e:1b:de', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'c7c7f16b-a2e3-4d1f-9b7e-a44a9ebb589c', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '9c52b1e4-82d2-4449-91f3-a624596bdc79', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1004.957076] env[62753]: DEBUG oslo.service.loopingcall [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1004.960262] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d20e083-2959-453a-8875-47955bc02613] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1004.960501] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9a565560-b951-4d2c-808f-453fbf8f7d36 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.979863] env[62753]: DEBUG oslo_vmware.api [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332615, 'name': PowerOffVM_Task, 'duration_secs': 0.205911} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1004.981128] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1004.981322] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1004.981572] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1004.981572] env[62753]: value = "task-1332618" [ 1004.981572] env[62753]: _type = "Task" [ 1004.981572] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1004.981761] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9892b9de-8fcf-40ef-af16-3f1cf215596f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1004.992469] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332618, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.032831] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332616, 'name': RemoveSnapshot_Task} progress is 80%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.058338] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1005.058492] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1005.058760] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Deleting the datastore file [datastore2] 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1005.058963] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a5d4d13c-b9ed-4a31-bee6-87cfa1370ae6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.065902] env[62753]: DEBUG oslo_vmware.api [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 1005.065902] env[62753]: value = "task-1332620" [ 1005.065902] env[62753]: _type = "Task" [ 1005.065902] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.076015] env[62753]: DEBUG oslo_vmware.api [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332620, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.271130] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6b8e6fd7-2f4b-4155-bea1-9bbdc7b409ac tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "interface-c3679c66-2763-4a04-8d14-e103bd952798-2498c430-c268-4e1e-b331-0a3b05eeae0a" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 10.053s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.290406] env[62753]: DEBUG nova.compute.manager [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Found 3 images (rotation: 2) {{(pid=62753) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4554}} [ 1005.290660] env[62753]: DEBUG nova.compute.manager [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Rotating out 1 backups {{(pid=62753) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4562}} [ 1005.290886] env[62753]: DEBUG nova.compute.manager [None req-d79a6955-6db3-4159-8b01-aa6121d45ad4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Deleting image 03c3734b-05a6-4c37-9f4e-406198dd6413 {{(pid=62753) _rotate_backups /opt/stack/nova/nova/compute/manager.py:4567}} [ 1005.311850] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332617, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.319105] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.319382] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.319572] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1005.406353] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "interface-53a76eb0-8b0c-4e47-8387-75df4c0b191f-2498c430-c268-4e1e-b331-0a3b05eeae0a" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.406445] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "interface-53a76eb0-8b0c-4e47-8387-75df4c0b191f-2498c430-c268-4e1e-b331-0a3b05eeae0a" acquired by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.406792] env[62753]: DEBUG nova.objects.instance [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lazy-loading 'flavor' on Instance uuid 53a76eb0-8b0c-4e47-8387-75df4c0b191f {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1005.501645] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332618, 'name': CreateVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.533742] env[62753]: DEBUG oslo_vmware.api [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332616, 'name': RemoveSnapshot_Task, 'duration_secs': 0.978582} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.534246] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Deleted Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1005.580840] env[62753]: DEBUG oslo_vmware.api [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332620, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.422875} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.586397] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1005.586397] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1005.586397] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1005.586397] env[62753]: INFO nova.compute.manager [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1005.586397] env[62753]: DEBUG oslo.service.loopingcall [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1005.586397] env[62753]: DEBUG nova.compute.manager [-] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1005.586397] env[62753]: DEBUG nova.network.neutron [-] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1005.678687] env[62753]: DEBUG nova.network.neutron [req-61133bb0-f347-427a-83e0-eee0f5646543 req-ad2dc458-37a4-4803-b99d-6c5c14b6f6fe service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Updated VIF entry in instance network info cache for port 12ca481a-0cb3-465e-b04c-802ae3ce6438. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1005.679125] env[62753]: DEBUG nova.network.neutron [req-61133bb0-f347-427a-83e0-eee0f5646543 req-ad2dc458-37a4-4803-b99d-6c5c14b6f6fe service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Updating instance_info_cache with network_info: [{"id": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "address": "fa:16:3e:57:5d:75", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12ca481a-0c", "ovs_interfaceid": "12ca481a-0cb3-465e-b04c-802ae3ce6438", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1005.689123] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "b5877d39-2f94-4a01-9db1-89a3e612095c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1005.689442] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "b5877d39-2f94-4a01-9db1-89a3e612095c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.003s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1005.806250] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332617, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.598557} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1005.806250] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 9d7066e1-3721-42d2-8e80-91eacbbd8b80/9d7066e1-3721-42d2-8e80-91eacbbd8b80.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1005.806554] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1005.806895] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f5d43204-3e90-464c-b5db-273bdf840232 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1005.814773] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1005.814773] env[62753]: value = "task-1332621" [ 1005.814773] env[62753]: _type = "Task" [ 1005.814773] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1005.832667] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332621, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1005.859138] env[62753]: DEBUG nova.network.neutron [req-5740a8a1-ccfc-437f-a7a3-c518777cf675 req-ab3db403-0342-4c98-9d1b-f2296b7f4ff5 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Updated VIF entry in instance network info cache for port 9c52b1e4-82d2-4449-91f3-a624596bdc79. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1005.859498] env[62753]: DEBUG nova.network.neutron [req-5740a8a1-ccfc-437f-a7a3-c518777cf675 req-ab3db403-0342-4c98-9d1b-f2296b7f4ff5 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Updating instance_info_cache with network_info: [{"id": "9c52b1e4-82d2-4449-91f3-a624596bdc79", "address": "fa:16:3e:8e:1b:de", "network": {"id": "12446a5e-7329-4f8f-aa28-a8b9d78e7573", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1766107072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "371a5e8b16be4cffbdeb2139757684fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7c7f16b-a2e3-4d1f-9b7e-a44a9ebb589c", "external-id": "nsx-vlan-transportzone-772", "segmentation_id": 772, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c52b1e4-82", "ovs_interfaceid": "9c52b1e4-82d2-4449-91f3-a624596bdc79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.001824] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332618, 'name': CreateVM_Task, 'duration_secs': 0.642488} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.002322] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 4d20e083-2959-453a-8875-47955bc02613] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1006.002968] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.003184] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.003580] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1006.003843] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3e86281f-3911-407e-8356-89944dd65fbe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.009660] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1006.009977] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 1006.009977] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5224dc28-128a-ef64-f9fd-e37aa75c2426" [ 1006.009977] env[62753]: _type = "Task" [ 1006.009977] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.010333] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1006.010333] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62753) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1006.019828] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5224dc28-128a-ef64-f9fd-e37aa75c2426, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.041072] env[62753]: WARNING nova.compute.manager [None req-fc929eb8-094c-4c8a-98e3-312f51d62170 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Image not found during snapshot: nova.exception.ImageNotFound: Image 701eee2c-daa2-4a72-a40a-6c24d5ec6909 could not be found. [ 1006.056568] env[62753]: DEBUG nova.objects.instance [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lazy-loading 'pci_requests' on Instance uuid 53a76eb0-8b0c-4e47-8387-75df4c0b191f {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1006.182525] env[62753]: DEBUG oslo_concurrency.lockutils [req-61133bb0-f347-427a-83e0-eee0f5646543 req-ad2dc458-37a4-4803-b99d-6c5c14b6f6fe service nova] Releasing lock "refresh_cache-c3679c66-2763-4a04-8d14-e103bd952798" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.195270] env[62753]: DEBUG nova.compute.manager [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1006.325995] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332621, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.070038} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.326303] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1006.327117] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8781518-d725-49c9-8489-ec7a7508732d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.351913] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 9d7066e1-3721-42d2-8e80-91eacbbd8b80/9d7066e1-3721-42d2-8e80-91eacbbd8b80.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1006.352224] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bc5fda2b-7c4f-4668-9c7d-33574186ad8b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.366873] env[62753]: DEBUG oslo_concurrency.lockutils [req-5740a8a1-ccfc-437f-a7a3-c518777cf675 req-ab3db403-0342-4c98-9d1b-f2296b7f4ff5 service nova] Releasing lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.373469] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1006.373469] env[62753]: value = "task-1332622" [ 1006.373469] env[62753]: _type = "Task" [ 1006.373469] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.381851] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332622, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.394187] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "refresh_cache-2b90824a-5d49-4ec9-bcf6-4f868ca876cf" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.394503] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "refresh_cache-2b90824a-5d49-4ec9-bcf6-4f868ca876cf" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.394779] env[62753]: DEBUG nova.network.neutron [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1006.501980] env[62753]: DEBUG nova.network.neutron [-] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1006.521214] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1006.521485] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Processing image 3b2f7019-a7ab-4e0e-9ea9-07c8a7334614 {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1006.521768] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.521872] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquired lock "[datastore2] devstack-image-cache_base/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.522073] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1006.522376] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-eac56f92-a40e-4bb1-bbde-29758ff10fd3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.531140] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1006.531360] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1006.532445] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c4c757f1-4ce6-4461-9ea9-3b14039a6500 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.538222] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 1006.538222] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f3ae5a-36aa-d9c2-9823-d4fa1e19adb4" [ 1006.538222] env[62753]: _type = "Task" [ 1006.538222] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.546758] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f3ae5a-36aa-d9c2-9823-d4fa1e19adb4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.558983] env[62753]: DEBUG nova.objects.base [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Object Instance<53a76eb0-8b0c-4e47-8387-75df4c0b191f> lazy-loaded attributes: flavor,pci_requests {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1006.559220] env[62753]: DEBUG nova.network.neutron [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1006.632257] env[62753]: DEBUG nova.policy [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9931b40cfd7846038805c6a4caedaac0', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b8e81660b30408c998e412f5fa81469', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1006.716255] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.716528] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.718120] env[62753]: INFO nova.compute.claims [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1006.799651] env[62753]: DEBUG nova.compute.manager [req-8d28ff17-669e-4c72-91a5-568fa9a4dddb req-7d4066a6-d46c-4403-b1bf-8a4411fb663a service nova] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Received event network-vif-deleted-f9490496-ad16-4da2-805c-a5187d087d67 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.884711] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332622, 'name': ReconfigVM_Task, 'duration_secs': 0.269521} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1006.885059] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 9d7066e1-3721-42d2-8e80-91eacbbd8b80/9d7066e1-3721-42d2-8e80-91eacbbd8b80.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1006.885765] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-c0fda1e1-648e-4891-bc9e-8e48bf8ed27b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.891674] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1006.891674] env[62753]: value = "task-1332623" [ 1006.891674] env[62753]: _type = "Task" [ 1006.891674] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.903188] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332623, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1006.903453] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "e1c59139-ff02-453e-b54f-a56b1e40fdb5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.903665] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "e1c59139-ff02-453e-b54f-a56b1e40fdb5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.903853] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "e1c59139-ff02-453e-b54f-a56b1e40fdb5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1006.904083] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "e1c59139-ff02-453e-b54f-a56b1e40fdb5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1006.904272] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "e1c59139-ff02-453e-b54f-a56b1e40fdb5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1006.907646] env[62753]: DEBUG nova.compute.manager [req-0c23146a-1b02-4ea8-b558-752495bda3f9 req-6bab5ab6-4f9f-4a79-8cf4-6c306ec68f80 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Received event network-changed-0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1006.907835] env[62753]: DEBUG nova.compute.manager [req-0c23146a-1b02-4ea8-b558-752495bda3f9 req-6bab5ab6-4f9f-4a79-8cf4-6c306ec68f80 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Refreshing instance network info cache due to event network-changed-0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1006.908048] env[62753]: DEBUG oslo_concurrency.lockutils [req-0c23146a-1b02-4ea8-b558-752495bda3f9 req-6bab5ab6-4f9f-4a79-8cf4-6c306ec68f80 service nova] Acquiring lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1006.908196] env[62753]: DEBUG oslo_concurrency.lockutils [req-0c23146a-1b02-4ea8-b558-752495bda3f9 req-6bab5ab6-4f9f-4a79-8cf4-6c306ec68f80 service nova] Acquired lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1006.908361] env[62753]: DEBUG nova.network.neutron [req-0c23146a-1b02-4ea8-b558-752495bda3f9 req-6bab5ab6-4f9f-4a79-8cf4-6c306ec68f80 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Refreshing network info cache for port 0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1006.909496] env[62753]: INFO nova.compute.manager [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Terminating instance [ 1006.913950] env[62753]: DEBUG nova.compute.manager [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1006.914126] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1006.915059] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48572fae-f1f0-44ca-b534-b7ea5c8e6c9d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.923843] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1006.924133] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f324b629-a7de-4f5f-a4ff-4310897cb773 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1006.929702] env[62753]: DEBUG oslo_vmware.api [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 1006.929702] env[62753]: value = "task-1332624" [ 1006.929702] env[62753]: _type = "Task" [ 1006.929702] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1006.938312] env[62753]: DEBUG oslo_vmware.api [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332624, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.004487] env[62753]: INFO nova.compute.manager [-] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Took 1.42 seconds to deallocate network for instance. [ 1007.009992] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1007.010155] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1007.051501] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Preparing fetch location {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1007.051761] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Fetch image to [datastore2] OSTACK_IMG_cd601da2-b3bd-4589-b4d4-21bbf7db9077/OSTACK_IMG_cd601da2-b3bd-4589-b4d4-21bbf7db9077.vmdk {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1007.052038] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Downloading stream optimized image 3b2f7019-a7ab-4e0e-9ea9-07c8a7334614 to [datastore2] OSTACK_IMG_cd601da2-b3bd-4589-b4d4-21bbf7db9077/OSTACK_IMG_cd601da2-b3bd-4589-b4d4-21bbf7db9077.vmdk on the data store datastore2 as vApp {{(pid=62753) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1007.052175] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Downloading image file data 3b2f7019-a7ab-4e0e-9ea9-07c8a7334614 to the ESX as VM named 'OSTACK_IMG_cd601da2-b3bd-4589-b4d4-21bbf7db9077' {{(pid=62753) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1007.128931] env[62753]: DEBUG nova.network.neutron [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Updating instance_info_cache with network_info: [{"id": "c86749ca-c0b7-4c20-a32a-fed94bb47e4d", "address": "fa:16:3e:8e:d2:8a", "network": {"id": "2823647f-d32d-4a1c-9e02-7577016a260e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1594498922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6c8a6bc946a469fa85cc3ab80d7333a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc86749ca-c0", "ovs_interfaceid": "c86749ca-c0b7-4c20-a32a-fed94bb47e4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.135644] env[62753]: DEBUG oslo_vmware.rw_handles [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1007.135644] env[62753]: value = "resgroup-9" [ 1007.135644] env[62753]: _type = "ResourcePool" [ 1007.135644] env[62753]: }. {{(pid=62753) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1007.136030] env[62753]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-7fd333b8-5089-442e-9708-e92d839d4a1a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.161485] env[62753]: DEBUG oslo_vmware.rw_handles [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lease: (returnval){ [ 1007.161485] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5215df93-8fef-cdb5-1f90-15590936d34c" [ 1007.161485] env[62753]: _type = "HttpNfcLease" [ 1007.161485] env[62753]: } obtained for vApp import into resource pool (val){ [ 1007.161485] env[62753]: value = "resgroup-9" [ 1007.161485] env[62753]: _type = "ResourcePool" [ 1007.161485] env[62753]: }. {{(pid=62753) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1007.162032] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the lease: (returnval){ [ 1007.162032] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5215df93-8fef-cdb5-1f90-15590936d34c" [ 1007.162032] env[62753]: _type = "HttpNfcLease" [ 1007.162032] env[62753]: } to be ready. {{(pid=62753) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1007.167645] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1007.167645] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5215df93-8fef-cdb5-1f90-15590936d34c" [ 1007.167645] env[62753]: _type = "HttpNfcLease" [ 1007.167645] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1007.404362] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332623, 'name': Rename_Task, 'duration_secs': 0.138284} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.404678] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1007.404964] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d7522880-85f4-4014-8ac9-b44488fdee02 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.413272] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1007.413272] env[62753]: value = "task-1332626" [ 1007.413272] env[62753]: _type = "Task" [ 1007.413272] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.424649] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332626, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.443081] env[62753]: DEBUG oslo_vmware.api [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332624, 'name': PowerOffVM_Task, 'duration_secs': 0.210155} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.443329] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1007.443474] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1007.443720] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b6cb6727-a455-4268-af71-02d724de0ab3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.511467] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1007.530815] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1007.531065] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1007.531234] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Deleting the datastore file [datastore2] e1c59139-ff02-453e-b54f-a56b1e40fdb5 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1007.531496] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c63443c8-e72d-414f-8356-71ffe7f1032a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.537661] env[62753]: DEBUG oslo_vmware.api [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for the task: (returnval){ [ 1007.537661] env[62753]: value = "task-1332628" [ 1007.537661] env[62753]: _type = "Task" [ 1007.537661] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1007.546393] env[62753]: DEBUG oslo_vmware.api [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332628, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1007.631772] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "refresh_cache-2b90824a-5d49-4ec9-bcf6-4f868ca876cf" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1007.653148] env[62753]: DEBUG nova.network.neutron [req-0c23146a-1b02-4ea8-b558-752495bda3f9 req-6bab5ab6-4f9f-4a79-8cf4-6c306ec68f80 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Updated VIF entry in instance network info cache for port 0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1007.653470] env[62753]: DEBUG nova.network.neutron [req-0c23146a-1b02-4ea8-b558-752495bda3f9 req-6bab5ab6-4f9f-4a79-8cf4-6c306ec68f80 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Updating instance_info_cache with network_info: [{"id": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "address": "fa:16:3e:6b:34:1e", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cd05bfa-82", "ovs_interfaceid": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1007.670683] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1007.670683] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5215df93-8fef-cdb5-1f90-15590936d34c" [ 1007.670683] env[62753]: _type = "HttpNfcLease" [ 1007.670683] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1007.923273] env[62753]: DEBUG oslo_vmware.api [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332626, 'name': PowerOnVM_Task, 'duration_secs': 0.487547} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1007.924532] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1007.924745] env[62753]: INFO nova.compute.manager [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Took 7.53 seconds to spawn the instance on the hypervisor. [ 1007.924929] env[62753]: DEBUG nova.compute.manager [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1007.925710] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-767f6475-eb94-482b-b7c9-8a91f651544e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.928730] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a1d5ce9-2569-4728-8577-26da5edc9b85 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.937327] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a2e437ef-67a5-4e18-989e-c037253efd43 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.973768] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b16e1027-2904-4398-84ef-40785d6e9bcb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.982009] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-578591b4-0f58-4b17-afea-f71fbe871921 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1007.995739] env[62753]: DEBUG nova.compute.provider_tree [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1008.005988] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1008.046772] env[62753]: DEBUG oslo_vmware.api [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Task: {'id': task-1332628, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.173967} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1008.047041] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1008.047233] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1008.047453] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1008.047585] env[62753]: INFO nova.compute.manager [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1008.047829] env[62753]: DEBUG oslo.service.loopingcall [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1008.048028] env[62753]: DEBUG nova.compute.manager [-] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1008.048122] env[62753]: DEBUG nova.network.neutron [-] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1008.155950] env[62753]: DEBUG oslo_concurrency.lockutils [req-0c23146a-1b02-4ea8-b558-752495bda3f9 req-6bab5ab6-4f9f-4a79-8cf4-6c306ec68f80 service nova] Releasing lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1008.158571] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81b40d24-7b77-4a8f-917b-4b1a12c9ea53 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.182618] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71fdcb66-c22f-45a5-827b-030928d2359f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.193899] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1008.193899] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5215df93-8fef-cdb5-1f90-15590936d34c" [ 1008.193899] env[62753]: _type = "HttpNfcLease" [ 1008.193899] env[62753]: } is ready. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1008.196338] env[62753]: DEBUG oslo_vmware.rw_handles [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1008.196338] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5215df93-8fef-cdb5-1f90-15590936d34c" [ 1008.196338] env[62753]: _type = "HttpNfcLease" [ 1008.196338] env[62753]: }. {{(pid=62753) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1008.196720] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Updating instance '2b90824a-5d49-4ec9-bcf6-4f868ca876cf' progress to 83 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1008.200698] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18b33748-84ac-4bfd-8da1-bcefdc03c8b3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.208035] env[62753]: DEBUG oslo_vmware.rw_handles [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Found VMDK URL: https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528f042b-523a-060e-46e1-047e61c1588f/disk-0.vmdk from lease info. {{(pid=62753) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1008.208233] env[62753]: DEBUG oslo_vmware.rw_handles [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Creating HTTP connection to write to file with size = 31664640 and URL = https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528f042b-523a-060e-46e1-047e61c1588f/disk-0.vmdk. {{(pid=62753) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1008.211718] env[62753]: DEBUG nova.network.neutron [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Successfully updated port: 2498c430-c268-4e1e-b331-0a3b05eeae0a {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1008.277728] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-4924d146-709a-47a3-a9fa-5df9676a54ba {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.315338] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b80027d0-85fc-4dfd-a81c-951c281e07d9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "c73b1ae0-dc98-47f7-babf-e96169384785" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.315629] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b80027d0-85fc-4dfd-a81c-951c281e07d9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "c73b1ae0-dc98-47f7-babf-e96169384785" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.315833] env[62753]: DEBUG nova.compute.manager [None req-b80027d0-85fc-4dfd-a81c-951c281e07d9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1008.316766] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-476d15fe-e423-42eb-9ebd-7d0268a065a3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.324048] env[62753]: DEBUG nova.compute.manager [None req-b80027d0-85fc-4dfd-a81c-951c281e07d9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62753) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1008.324723] env[62753]: DEBUG nova.objects.instance [None req-b80027d0-85fc-4dfd-a81c-951c281e07d9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lazy-loading 'flavor' on Instance uuid c73b1ae0-dc98-47f7-babf-e96169384785 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1008.450237] env[62753]: INFO nova.compute.manager [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Took 20.74 seconds to build instance. [ 1008.499099] env[62753]: DEBUG nova.scheduler.client.report [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1008.514120] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1008.705310] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1008.706654] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f34c5ca8-48e8-4e8a-a610-a4f4ae4cb679 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.714277] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.714499] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.714692] env[62753]: DEBUG nova.network.neutron [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1008.715998] env[62753]: DEBUG oslo_vmware.api [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 1008.715998] env[62753]: value = "task-1332629" [ 1008.715998] env[62753]: _type = "Task" [ 1008.715998] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.727095] env[62753]: DEBUG oslo_vmware.api [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332629, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.829869] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b80027d0-85fc-4dfd-a81c-951c281e07d9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1008.830152] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d13b61ad-890d-4847-85ff-ea2e15a01873 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1008.837589] env[62753]: DEBUG oslo_vmware.api [None req-b80027d0-85fc-4dfd-a81c-951c281e07d9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1008.837589] env[62753]: value = "task-1332630" [ 1008.837589] env[62753]: _type = "Task" [ 1008.837589] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1008.847087] env[62753]: DEBUG oslo_vmware.api [None req-b80027d0-85fc-4dfd-a81c-951c281e07d9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332630, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1008.862883] env[62753]: DEBUG nova.network.neutron [-] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1008.880632] env[62753]: DEBUG nova.compute.manager [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Received event network-vif-plugged-2498c430-c268-4e1e-b331-0a3b05eeae0a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.880968] env[62753]: DEBUG oslo_concurrency.lockutils [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] Acquiring lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1008.881271] env[62753]: DEBUG oslo_concurrency.lockutils [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] Lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1008.881526] env[62753]: DEBUG oslo_concurrency.lockutils [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] Lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1008.881819] env[62753]: DEBUG nova.compute.manager [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] No waiting events found dispatching network-vif-plugged-2498c430-c268-4e1e-b331-0a3b05eeae0a {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1008.882064] env[62753]: WARNING nova.compute.manager [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Received unexpected event network-vif-plugged-2498c430-c268-4e1e-b331-0a3b05eeae0a for instance with vm_state active and task_state None. [ 1008.882248] env[62753]: DEBUG nova.compute.manager [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Received event network-changed-2498c430-c268-4e1e-b331-0a3b05eeae0a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1008.882439] env[62753]: DEBUG nova.compute.manager [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Refreshing instance network info cache due to event network-changed-2498c430-c268-4e1e-b331-0a3b05eeae0a. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1008.882612] env[62753]: DEBUG oslo_concurrency.lockutils [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] Acquiring lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.885532] env[62753]: INFO nova.compute.manager [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Rescuing [ 1008.885798] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "refresh_cache-9d7066e1-3721-42d2-8e80-91eacbbd8b80" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1008.885954] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquired lock "refresh_cache-9d7066e1-3721-42d2-8e80-91eacbbd8b80" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1008.886131] env[62753]: DEBUG nova.network.neutron [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1008.952955] env[62753]: DEBUG oslo_concurrency.lockutils [None req-84907e66-f004-439d-89d7-c61c73a02c68 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "9d7066e1-3721-42d2-8e80-91eacbbd8b80" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.252s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.007018] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.290s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.007742] env[62753]: DEBUG nova.compute.manager [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1009.012455] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.501s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1009.012751] env[62753]: DEBUG nova.objects.instance [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lazy-loading 'resources' on Instance uuid 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1009.013943] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.014617] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1009.228745] env[62753]: DEBUG oslo_vmware.api [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332629, 'name': PowerOnVM_Task, 'duration_secs': 0.487446} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.229074] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1009.229355] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d46ed841-8e56-409d-ab2c-7567a81eee45 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Updating instance '2b90824a-5d49-4ec9-bcf6-4f868ca876cf' progress to 100 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1009.255621] env[62753]: WARNING nova.network.neutron [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] 8a9463a1-a861-4ade-8483-640392be0a89 already exists in list: networks containing: ['8a9463a1-a861-4ade-8483-640392be0a89']. ignoring it [ 1009.347587] env[62753]: DEBUG oslo_vmware.api [None req-b80027d0-85fc-4dfd-a81c-951c281e07d9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332630, 'name': PowerOffVM_Task, 'duration_secs': 0.175745} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1009.350303] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b80027d0-85fc-4dfd-a81c-951c281e07d9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1009.350492] env[62753]: DEBUG nova.compute.manager [None req-b80027d0-85fc-4dfd-a81c-951c281e07d9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1009.352511] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d90646a7-b05e-4f82-903f-0bd43f28a279 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.366257] env[62753]: INFO nova.compute.manager [-] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Took 1.32 seconds to deallocate network for instance. [ 1009.377720] env[62753]: DEBUG oslo_vmware.rw_handles [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Completed reading data from the image iterator. {{(pid=62753) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1009.377894] env[62753]: DEBUG oslo_vmware.rw_handles [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Getting lease state for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528f042b-523a-060e-46e1-047e61c1588f/disk-0.vmdk. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1009.378760] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17713485-09d3-4abe-9f29-2d008467ceb7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.385555] env[62753]: DEBUG oslo_vmware.rw_handles [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528f042b-523a-060e-46e1-047e61c1588f/disk-0.vmdk is in state: ready. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1009.385748] env[62753]: DEBUG oslo_vmware.rw_handles [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Releasing lease for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528f042b-523a-060e-46e1-047e61c1588f/disk-0.vmdk. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1009.386011] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-4fc054ce-38fe-4016-8215-cc071636eb6a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.516055] env[62753]: DEBUG nova.compute.utils [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1009.523075] env[62753]: DEBUG nova.compute.manager [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1009.523075] env[62753]: DEBUG nova.network.neutron [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1009.525383] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.574768] env[62753]: DEBUG oslo_vmware.rw_handles [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Closed VMDK write handle for https://esx7c1n1.openstack.eu-de-1.cloud.sap/nfc/528f042b-523a-060e-46e1-047e61c1588f/disk-0.vmdk. {{(pid=62753) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1009.575049] env[62753]: INFO nova.virt.vmwareapi.images [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Downloaded image file data 3b2f7019-a7ab-4e0e-9ea9-07c8a7334614 [ 1009.575892] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea0914f3-c2a8-4b9c-87e6-49efe9b0605e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.600998] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-eb78046b-ff9d-4458-99d2-54ff485d5cfa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.621388] env[62753]: DEBUG nova.policy [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae52823f7c93454e8089261c3ca44321', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d52d6eaee934be5ab0e0003df1ce316', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1009.652389] env[62753]: INFO nova.virt.vmwareapi.images [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] The imported VM was unregistered [ 1009.654865] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Caching image {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1009.655656] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Creating directory with path [datastore2] devstack-image-cache_base/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614 {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1009.655943] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95beb15f-7ae9-49ee-bc58-b7509677c207 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.676482] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Created directory with path [datastore2] devstack-image-cache_base/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614 {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1009.676695] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_cd601da2-b3bd-4589-b4d4-21bbf7db9077/OSTACK_IMG_cd601da2-b3bd-4589-b4d4-21bbf7db9077.vmdk to [datastore2] devstack-image-cache_base/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614.vmdk. {{(pid=62753) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1009.676953] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-fd05fb60-7c06-44d2-8a8a-180ce2f3aef9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.685343] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 1009.685343] env[62753]: value = "task-1332632" [ 1009.685343] env[62753]: _type = "Task" [ 1009.685343] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1009.696755] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332632, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1009.713701] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71925801-b408-47db-8398-6b20ec166ebf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.720421] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e7cbfc-cc8b-4aac-9b22-bf4f41497fca {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.752533] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-80c68f59-0c54-464c-8148-191ca7e5c54b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.760386] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7630b9a0-e254-492e-9641-5483bbe793d4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1009.773447] env[62753]: DEBUG nova.compute.provider_tree [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1009.778792] env[62753]: DEBUG nova.network.neutron [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Updating instance_info_cache with network_info: [{"id": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "address": "fa:16:3e:6b:34:1e", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cd05bfa-82", "ovs_interfaceid": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2498c430-c268-4e1e-b331-0a3b05eeae0a", "address": "fa:16:3e:f3:a1:e8", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2498c430-c2", "ovs_interfaceid": "2498c430-c268-4e1e-b331-0a3b05eeae0a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.785146] env[62753]: DEBUG nova.network.neutron [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Updating instance_info_cache with network_info: [{"id": "7f800ba5-2f01-46b7-b84a-88f900822bf2", "address": "fa:16:3e:03:90:c5", "network": {"id": "22a90636-182f-4944-893f-329f446b07c4", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1513076888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "37368e15fa5a4a0a97ac1c22fcf0bb80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f800ba5-2f", "ovs_interfaceid": "7f800ba5-2f01-46b7-b84a-88f900822bf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1009.864503] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b80027d0-85fc-4dfd-a81c-951c281e07d9 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "c73b1ae0-dc98-47f7-babf-e96169384785" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.549s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1009.874528] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1009.984405] env[62753]: DEBUG nova.network.neutron [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Successfully created port: 96621c20-3355-4040-ad41-85fbfded9580 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1010.022844] env[62753]: DEBUG nova.compute.manager [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1010.197590] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332632, 'name': MoveVirtualDisk_Task} progress is 9%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.279708] env[62753]: DEBUG nova.scheduler.client.report [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1010.283383] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.283704] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1010.283864] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.284484] env[62753]: DEBUG oslo_concurrency.lockutils [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] Acquired lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1010.284649] env[62753]: DEBUG nova.network.neutron [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Refreshing network info cache for port 2498c430-c268-4e1e-b331-0a3b05eeae0a {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1010.286475] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-494496af-e9bc-4265-a395-b2aa58cd5d3c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.291964] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Releasing lock "refresh_cache-9d7066e1-3721-42d2-8e80-91eacbbd8b80" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1010.312450] env[62753]: DEBUG nova.virt.hardware [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=,min_ram=,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1010.312717] env[62753]: DEBUG nova.virt.hardware [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1010.312880] env[62753]: DEBUG nova.virt.hardware [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1010.313089] env[62753]: DEBUG nova.virt.hardware [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1010.313248] env[62753]: DEBUG nova.virt.hardware [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1010.313402] env[62753]: DEBUG nova.virt.hardware [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1010.313613] env[62753]: DEBUG nova.virt.hardware [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1010.313779] env[62753]: DEBUG nova.virt.hardware [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1010.313964] env[62753]: DEBUG nova.virt.hardware [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1010.314180] env[62753]: DEBUG nova.virt.hardware [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1010.314366] env[62753]: DEBUG nova.virt.hardware [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1010.320784] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Reconfiguring VM to attach interface {{(pid=62753) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1923}} [ 1010.323434] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-32e3ee60-e7dd-4221-9682-b6384330a682 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.338762] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.339031] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.339245] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.339437] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.339884] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.346797] env[62753]: INFO nova.compute.manager [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Terminating instance [ 1010.348328] env[62753]: DEBUG oslo_vmware.api [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 1010.348328] env[62753]: value = "task-1332633" [ 1010.348328] env[62753]: _type = "Task" [ 1010.348328] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.348861] env[62753]: DEBUG nova.compute.manager [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1010.349108] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1010.350027] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e7b8d826-c727-4451-b2ab-764c248b9548 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.361557] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1010.364987] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-333c5cbd-8c9f-47d3-b095-949bca4d5151 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.366786] env[62753]: DEBUG oslo_vmware.api [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332633, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.371695] env[62753]: DEBUG oslo_vmware.api [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 1010.371695] env[62753]: value = "task-1332634" [ 1010.371695] env[62753]: _type = "Task" [ 1010.371695] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.380494] env[62753]: DEBUG oslo_vmware.api [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332634, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.699169] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332632, 'name': MoveVirtualDisk_Task} progress is 21%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.785495] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.773s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.789031] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.263s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.789031] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1010.789031] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62753) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1010.789361] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.915s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1010.789537] env[62753]: DEBUG nova.objects.instance [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lazy-loading 'resources' on Instance uuid e1c59139-ff02-453e-b54f-a56b1e40fdb5 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1010.792081] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-224d3fac-4b6c-4d77-803f-7d33b26dffbc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.805886] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d9a3614-d9e5-4e13-8fb0-0be4fe2ac381 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.811461] env[62753]: INFO nova.scheduler.client.report [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Deleted allocations for instance 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99 [ 1010.826359] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23bdc3b3-f649-4678-96bc-5d6d52eddb90 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.834063] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a745af15-a0ae-41c6-a969-c04c129400d6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.871215] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1010.871740] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180685MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=62753) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1010.871923] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1010.872956] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-85086c53-5409-4d3d-a958-71992290ef10 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.892008] env[62753]: DEBUG oslo_vmware.api [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332633, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.892639] env[62753]: DEBUG oslo_vmware.api [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332634, 'name': PowerOffVM_Task, 'duration_secs': 0.191298} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1010.894051] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1010.894257] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1010.894582] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1010.894582] env[62753]: value = "task-1332635" [ 1010.894582] env[62753]: _type = "Task" [ 1010.894582] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.894788] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-79e4621a-b024-40c1-aa18-5fef3072b903 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.905740] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332635, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1010.975171] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1010.975421] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1010.975659] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Deleting the datastore file [datastore1] 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1010.975958] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a30aa692-1238-406b-b974-3192945cbc56 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1010.983567] env[62753]: DEBUG oslo_vmware.api [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for the task: (returnval){ [ 1010.983567] env[62753]: value = "task-1332637" [ 1010.983567] env[62753]: _type = "Task" [ 1010.983567] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1010.992625] env[62753]: DEBUG oslo_vmware.api [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332637, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.032871] env[62753]: DEBUG nova.compute.manager [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1011.061074] env[62753]: DEBUG nova.virt.hardware [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1011.061332] env[62753]: DEBUG nova.virt.hardware [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1011.061482] env[62753]: DEBUG nova.virt.hardware [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1011.061659] env[62753]: DEBUG nova.virt.hardware [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1011.061807] env[62753]: DEBUG nova.virt.hardware [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1011.061970] env[62753]: DEBUG nova.virt.hardware [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1011.062226] env[62753]: DEBUG nova.virt.hardware [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1011.062391] env[62753]: DEBUG nova.virt.hardware [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1011.062561] env[62753]: DEBUG nova.virt.hardware [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1011.062723] env[62753]: DEBUG nova.virt.hardware [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1011.062894] env[62753]: DEBUG nova.virt.hardware [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1011.064273] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5f4f8b3-d561-4018-bcf4-a17eed595d11 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.073338] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4fddff2-4ae2-4bb6-81fd-335c2b31313a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.198594] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332632, 'name': MoveVirtualDisk_Task} progress is 43%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.306226] env[62753]: DEBUG nova.network.neutron [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Updated VIF entry in instance network info cache for port 2498c430-c268-4e1e-b331-0a3b05eeae0a. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1011.306811] env[62753]: DEBUG nova.network.neutron [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Updating instance_info_cache with network_info: [{"id": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "address": "fa:16:3e:6b:34:1e", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cd05bfa-82", "ovs_interfaceid": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}, {"id": "2498c430-c268-4e1e-b331-0a3b05eeae0a", "address": "fa:16:3e:f3:a1:e8", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2498c430-c2", "ovs_interfaceid": "2498c430-c268-4e1e-b331-0a3b05eeae0a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": true, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.320731] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1b19508f-16c5-456f-b8a2-27a8791a95af tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "3009b1e4-53b1-4b12-8aaf-3c36b37c5d99" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.895s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.383136] env[62753]: DEBUG oslo_vmware.api [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332633, 'name': ReconfigVM_Task, 'duration_secs': 0.683499} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.383673] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.383888] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Reconfigured VM to attach interface {{(pid=62753) attach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1937}} [ 1011.407399] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332635, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.497159] env[62753]: DEBUG oslo_vmware.api [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Task: {'id': task-1332637, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.334655} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1011.497496] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1011.497736] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1011.497974] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1011.498190] env[62753]: INFO nova.compute.manager [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1011.498480] env[62753]: DEBUG oslo.service.loopingcall [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1011.498713] env[62753]: DEBUG nova.compute.manager [-] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1011.498837] env[62753]: DEBUG nova.network.neutron [-] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1011.504713] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e73e5ab-f32b-4195-a919-dfaccee73627 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.513289] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-378c5685-4714-4e86-bca9-2f3fd713fd11 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.550801] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4144e84e-6c5e-4f66-8e30-438e6d1d1843 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.561303] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3638f63f-3edf-4fe8-ab7d-71a119d8a9b6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1011.578805] env[62753]: DEBUG nova.compute.provider_tree [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1011.659407] env[62753]: DEBUG nova.compute.manager [req-3a7c1032-1963-4ac5-a318-3adee9d2b8c5 req-017eb788-9a2f-4714-bd4f-d6c730e829af service nova] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Received event network-vif-plugged-96621c20-3355-4040-ad41-85fbfded9580 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.659668] env[62753]: DEBUG oslo_concurrency.lockutils [req-3a7c1032-1963-4ac5-a318-3adee9d2b8c5 req-017eb788-9a2f-4714-bd4f-d6c730e829af service nova] Acquiring lock "b5877d39-2f94-4a01-9db1-89a3e612095c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.659891] env[62753]: DEBUG oslo_concurrency.lockutils [req-3a7c1032-1963-4ac5-a318-3adee9d2b8c5 req-017eb788-9a2f-4714-bd4f-d6c730e829af service nova] Lock "b5877d39-2f94-4a01-9db1-89a3e612095c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.660077] env[62753]: DEBUG oslo_concurrency.lockutils [req-3a7c1032-1963-4ac5-a318-3adee9d2b8c5 req-017eb788-9a2f-4714-bd4f-d6c730e829af service nova] Lock "b5877d39-2f94-4a01-9db1-89a3e612095c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.660267] env[62753]: DEBUG nova.compute.manager [req-3a7c1032-1963-4ac5-a318-3adee9d2b8c5 req-017eb788-9a2f-4714-bd4f-d6c730e829af service nova] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] No waiting events found dispatching network-vif-plugged-96621c20-3355-4040-ad41-85fbfded9580 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1011.660527] env[62753]: WARNING nova.compute.manager [req-3a7c1032-1963-4ac5-a318-3adee9d2b8c5 req-017eb788-9a2f-4714-bd4f-d6c730e829af service nova] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Received unexpected event network-vif-plugged-96621c20-3355-4040-ad41-85fbfded9580 for instance with vm_state building and task_state spawning. [ 1011.702212] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332632, 'name': MoveVirtualDisk_Task} progress is 63%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1011.781876] env[62753]: DEBUG nova.compute.manager [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Stashing vm_state: stopped {{(pid=62753) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1011.813385] env[62753]: DEBUG oslo_concurrency.lockutils [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] Releasing lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1011.813659] env[62753]: DEBUG nova.compute.manager [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Received event network-vif-deleted-99cbc723-6896-4ee7-bc42-e8fdac96ea65 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1011.813841] env[62753]: INFO nova.compute.manager [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Neutron deleted interface 99cbc723-6896-4ee7-bc42-e8fdac96ea65; detaching it from the instance and deleting it from the info cache [ 1011.814178] env[62753]: DEBUG nova.network.neutron [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1011.889721] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1cb654f9-a1e5-49db-8034-217687718728 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "interface-53a76eb0-8b0c-4e47-8387-75df4c0b191f-2498c430-c268-4e1e-b331-0a3b05eeae0a" "released" by "nova.compute.manager.ComputeManager.attach_interface..do_attach_interface" :: held 6.483s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1011.896811] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1011.897105] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1011.897315] env[62753]: DEBUG nova.compute.manager [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Going to confirm migration 2 {{(pid=62753) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1011.908163] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332635, 'name': PowerOffVM_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.029593] env[62753]: DEBUG nova.network.neutron [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Successfully updated port: 96621c20-3355-4040-ad41-85fbfded9580 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1012.082236] env[62753]: DEBUG nova.scheduler.client.report [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1012.202139] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332632, 'name': MoveVirtualDisk_Task} progress is 83%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.300202] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.318275] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-a9f58246-d9a2-4633-923b-c0579624a1d4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.328837] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-73bb38af-22d2-4e56-af92-6f024f76790b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.359839] env[62753]: DEBUG nova.compute.manager [req-73452c6a-96c4-4424-acd6-f3f5393a6e83 req-83a25e50-a81b-4e0a-9c37-756cb10bfde6 service nova] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Detach interface failed, port_id=99cbc723-6896-4ee7-bc42-e8fdac96ea65, reason: Instance e1c59139-ff02-453e-b54f-a56b1e40fdb5 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1012.410960] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332635, 'name': PowerOffVM_Task, 'duration_secs': 1.150411} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.413317] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1012.413943] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-339055db-f113-4ff7-b8e8-7fb35af23f2a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.437561] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-deb8d5ea-2a34-40fd-b573-b1476a97c1a5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.442045] env[62753]: DEBUG nova.compute.manager [req-2d798cc0-5c9d-4221-96db-4b3c8fd41611 req-001a7a6d-afa6-44ba-9a5b-5caf084e8907 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Received event network-vif-deleted-7fb046f4-e470-4f6d-8fa3-73f3b0157020 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1012.442045] env[62753]: INFO nova.compute.manager [req-2d798cc0-5c9d-4221-96db-4b3c8fd41611 req-001a7a6d-afa6-44ba-9a5b-5caf084e8907 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Neutron deleted interface 7fb046f4-e470-4f6d-8fa3-73f3b0157020; detaching it from the instance and deleting it from the info cache [ 1012.442234] env[62753]: DEBUG nova.network.neutron [req-2d798cc0-5c9d-4221-96db-4b3c8fd41611 req-001a7a6d-afa6-44ba-9a5b-5caf084e8907 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.477456] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1012.478577] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-34de0a5c-31d5-4fc9-886e-35ee990531e3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.481367] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "refresh_cache-2b90824a-5d49-4ec9-bcf6-4f868ca876cf" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.481617] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "refresh_cache-2b90824a-5d49-4ec9-bcf6-4f868ca876cf" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.481719] env[62753]: DEBUG nova.network.neutron [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1012.481939] env[62753]: DEBUG nova.objects.instance [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lazy-loading 'info_cache' on Instance uuid 2b90824a-5d49-4ec9-bcf6-4f868ca876cf {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1012.493018] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1012.493018] env[62753]: value = "task-1332638" [ 1012.493018] env[62753]: _type = "Task" [ 1012.493018] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.499748] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] VM already powered off {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1012.500170] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1012.500464] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.500663] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.500898] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1012.501187] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-e681d154-60f8-459f-a311-34ff15b11bce {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.516318] env[62753]: DEBUG nova.network.neutron [-] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1012.518968] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1012.519271] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1012.520353] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6ac139a3-b70b-49dd-9411-5bd01d9d0454 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.527579] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1012.527579] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5229dcb0-462e-8748-b883-774eb3b63408" [ 1012.527579] env[62753]: _type = "Task" [ 1012.527579] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.532131] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "refresh_cache-b5877d39-2f94-4a01-9db1-89a3e612095c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1012.532286] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "refresh_cache-b5877d39-2f94-4a01-9db1-89a3e612095c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1012.532423] env[62753]: DEBUG nova.network.neutron [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1012.539626] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5229dcb0-462e-8748-b883-774eb3b63408, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.592570] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.803s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1012.597828] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 1.726s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.599603] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "b1e4c8f9-471d-4cca-9a12-56191767a5b1" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1012.599889] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "b1e4c8f9-471d-4cca-9a12-56191767a5b1" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1012.617181] env[62753]: INFO nova.scheduler.client.report [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Deleted allocations for instance e1c59139-ff02-453e-b54f-a56b1e40fdb5 [ 1012.702165] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332632, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.973413} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1012.702462] env[62753]: INFO nova.virt.vmwareapi.ds_util [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_cd601da2-b3bd-4589-b4d4-21bbf7db9077/OSTACK_IMG_cd601da2-b3bd-4589-b4d4-21bbf7db9077.vmdk to [datastore2] devstack-image-cache_base/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614.vmdk. [ 1012.702706] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Cleaning up location [datastore2] OSTACK_IMG_cd601da2-b3bd-4589-b4d4-21bbf7db9077 {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1012.702838] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_cd601da2-b3bd-4589-b4d4-21bbf7db9077 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1012.703127] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-516443e9-d33a-4e9a-b37b-53f8968004c8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.709889] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 1012.709889] env[62753]: value = "task-1332639" [ 1012.709889] env[62753]: _type = "Task" [ 1012.709889] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1012.717879] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332639, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1012.945452] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-af4746b6-ae32-4364-bd04-f9eade6c2a7f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.956960] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a8740990-94d9-4c30-b452-82c08fe8d603 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1012.983333] env[62753]: DEBUG nova.compute.manager [req-2d798cc0-5c9d-4221-96db-4b3c8fd41611 req-001a7a6d-afa6-44ba-9a5b-5caf084e8907 service nova] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Detach interface failed, port_id=7fb046f4-e470-4f6d-8fa3-73f3b0157020, reason: Instance 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1013.021587] env[62753]: INFO nova.compute.manager [-] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Took 1.52 seconds to deallocate network for instance. [ 1013.041490] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5229dcb0-462e-8748-b883-774eb3b63408, 'name': SearchDatastore_Task, 'duration_secs': 0.04947} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.042623] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3034b9cb-a19e-4ed4-a7f1-f22e1998b8e3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.048742] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1013.048742] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dbcca6-ce8d-446a-ad1a-62b80d406175" [ 1013.048742] env[62753]: _type = "Task" [ 1013.048742] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.056933] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dbcca6-ce8d-446a-ad1a-62b80d406175, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.071350] env[62753]: DEBUG nova.network.neutron [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1013.108716] env[62753]: DEBUG nova.compute.manager [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1013.126243] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "interface-53a76eb0-8b0c-4e47-8387-75df4c0b191f-2498c430-c268-4e1e-b331-0a3b05eeae0a" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.126520] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "interface-53a76eb0-8b0c-4e47-8387-75df4c0b191f-2498c430-c268-4e1e-b331-0a3b05eeae0a" acquired by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1013.128339] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e764bec4-b8a8-41ac-a30d-6ffb2dc1f129 tempest-ImagesTestJSON-1896168473 tempest-ImagesTestJSON-1896168473-project-member] Lock "e1c59139-ff02-453e-b54f-a56b1e40fdb5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.225s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1013.216667] env[62753]: DEBUG nova.network.neutron [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Updating instance_info_cache with network_info: [{"id": "96621c20-3355-4040-ad41-85fbfded9580", "address": "fa:16:3e:a7:09:43", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96621c20-33", "ovs_interfaceid": "96621c20-3355-4040-ad41-85fbfded9580", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1013.221158] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332639, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.045344} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.221599] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1013.221766] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Releasing lock "[datastore2] devstack-image-cache_base/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.222082] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614.vmdk to [datastore2] 4d20e083-2959-453a-8875-47955bc02613/4d20e083-2959-453a-8875-47955bc02613.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1013.222346] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8c0ed2d3-bacf-4c8f-8509-69dcd5498786 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.228692] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 1013.228692] env[62753]: value = "task-1332640" [ 1013.228692] env[62753]: _type = "Task" [ 1013.228692] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.237260] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332640, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.528360] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.559576] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dbcca6-ce8d-446a-ad1a-62b80d406175, 'name': SearchDatastore_Task, 'duration_secs': 0.053842} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1013.559833] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.560102] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 9d7066e1-3721-42d2-8e80-91eacbbd8b80/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk. {{(pid=62753) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1013.560372] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3905c741-290c-4980-96aa-a6820ee353d1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.566649] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1013.566649] env[62753]: value = "task-1332641" [ 1013.566649] env[62753]: _type = "Task" [ 1013.566649] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.574257] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332641, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.612222] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Applying migration context for instance 2b90824a-5d49-4ec9-bcf6-4f868ca876cf as it has an incoming, in-progress migration 4a936932-16ef-44c3-898d-da10d29f220b. Migration status is confirming {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1013.613659] env[62753]: INFO nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Updating resource usage from migration 4a936932-16ef-44c3-898d-da10d29f220b [ 1013.613997] env[62753]: INFO nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating resource usage from migration 1c84581b-3dc0-4087-a0c2-fa39eb566788 [ 1013.630649] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1013.630839] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1013.631790] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23018b50-dd16-4eb7-af95-8f68b6044487 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.637898] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1013.639051] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance c3679c66-2763-4a04-8d14-e103bd952798 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1013.639193] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1013.639313] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 53a76eb0-8b0c-4e47-8387-75df4c0b191f actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1013.639429] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance bba5e3cb-c896-4ce7-ac7f-1cf2aba20465 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1013.639541] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 04fd0022-f20a-4217-9e47-5381635b17a6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1013.639654] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 4d20e083-2959-453a-8875-47955bc02613 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1013.639765] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Migration 4a936932-16ef-44c3-898d-da10d29f220b is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1013.639874] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 2b90824a-5d49-4ec9-bcf6-4f868ca876cf actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1013.639983] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 9d7066e1-3721-42d2-8e80-91eacbbd8b80 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1013.640112] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance b5877d39-2f94-4a01-9db1-89a3e612095c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1013.640220] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Migration 1c84581b-3dc0-4087-a0c2-fa39eb566788 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1013.640328] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance c73b1ae0-dc98-47f7-babf-e96169384785 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1013.659123] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48f2057d-a6a1-489f-a073-8133a62cec49 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.685163] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Reconfiguring VM to detach interface {{(pid=62753) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1968}} [ 1013.685725] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8a3e1683-614a-4650-a114-42298d73db37 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.703640] env[62753]: DEBUG oslo_vmware.api [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 1013.703640] env[62753]: value = "task-1332642" [ 1013.703640] env[62753]: _type = "Task" [ 1013.703640] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.714150] env[62753]: DEBUG oslo_vmware.api [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332642, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.722138] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "refresh_cache-b5877d39-2f94-4a01-9db1-89a3e612095c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1013.722503] env[62753]: DEBUG nova.compute.manager [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Instance network_info: |[{"id": "96621c20-3355-4040-ad41-85fbfded9580", "address": "fa:16:3e:a7:09:43", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96621c20-33", "ovs_interfaceid": "96621c20-3355-4040-ad41-85fbfded9580", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1013.722989] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:09:43', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a10c88d7-d13f-44fd-acee-7a734eb5f56a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '96621c20-3355-4040-ad41-85fbfded9580', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1013.730801] env[62753]: DEBUG oslo.service.loopingcall [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1013.731164] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1013.734726] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-dc57207c-089f-4630-9956-cbd4988c2b76 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1013.754029] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332640, 'name': CopyVirtualDisk_Task} progress is 9%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.755320] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1013.755320] env[62753]: value = "task-1332643" [ 1013.755320] env[62753]: _type = "Task" [ 1013.755320] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1013.762767] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332643, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1013.828110] env[62753]: DEBUG nova.network.neutron [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Updating instance_info_cache with network_info: [{"id": "c86749ca-c0b7-4c20-a32a-fed94bb47e4d", "address": "fa:16:3e:8e:d2:8a", "network": {"id": "2823647f-d32d-4a1c-9e02-7577016a260e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1594498922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6c8a6bc946a469fa85cc3ab80d7333a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapc86749ca-c0", "ovs_interfaceid": "c86749ca-c0b7-4c20-a32a-fed94bb47e4d", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1014.077615] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332641, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.143572] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance b1e4c8f9-471d-4cca-9a12-56191767a5b1 has been scheduled to this compute host, the scheduler has made an allocation against this compute node but the instance has yet to start. Skipping heal of allocation: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1764}} [ 1014.143941] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Total usable vcpus: 48, total allocated vcpus: 11 {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1014.144212] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2688MB phys_disk=200GB used_disk=11GB total_vcpus=48 used_vcpus=11 pci_stats=[] {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1014.214250] env[62753]: DEBUG oslo_vmware.api [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332642, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.246505] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332640, 'name': CopyVirtualDisk_Task} progress is 21%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.274372] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332643, 'name': CreateVM_Task, 'duration_secs': 0.337949} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.274682] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1014.275647] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.275647] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.275953] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1014.280012] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-f434054d-3d5c-48ba-a4d7-5b6c29f42a54 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.285365] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1014.285365] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5211fa2a-088d-909b-d9ac-101657136c26" [ 1014.285365] env[62753]: _type = "Task" [ 1014.285365] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.293980] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5211fa2a-088d-909b-d9ac-101657136c26, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.330737] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "refresh_cache-2b90824a-5d49-4ec9-bcf6-4f868ca876cf" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.331049] env[62753]: DEBUG nova.objects.instance [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lazy-loading 'migration_context' on Instance uuid 2b90824a-5d49-4ec9-bcf6-4f868ca876cf {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1014.384895] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-81e3330a-87b4-40d0-a404-c53dbd3be3b1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.394080] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ecd2bcb4-e80a-4ad4-8227-9db53b52964a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.429323] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39d65872-a65b-448d-a977-035acfff2867 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.437310] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ba6ea59-1ac9-4b72-84dc-7d80de1c5eff {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.451581] env[62753]: DEBUG nova.compute.provider_tree [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1014.458854] env[62753]: DEBUG nova.compute.manager [req-a212e5b0-7b62-4a99-8f14-11ab8ca9dac3 req-2582ef08-e74e-47de-af91-06b242db741f service nova] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Received event network-changed-96621c20-3355-4040-ad41-85fbfded9580 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1014.459468] env[62753]: DEBUG nova.compute.manager [req-a212e5b0-7b62-4a99-8f14-11ab8ca9dac3 req-2582ef08-e74e-47de-af91-06b242db741f service nova] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Refreshing instance network info cache due to event network-changed-96621c20-3355-4040-ad41-85fbfded9580. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1014.459468] env[62753]: DEBUG oslo_concurrency.lockutils [req-a212e5b0-7b62-4a99-8f14-11ab8ca9dac3 req-2582ef08-e74e-47de-af91-06b242db741f service nova] Acquiring lock "refresh_cache-b5877d39-2f94-4a01-9db1-89a3e612095c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.459468] env[62753]: DEBUG oslo_concurrency.lockutils [req-a212e5b0-7b62-4a99-8f14-11ab8ca9dac3 req-2582ef08-e74e-47de-af91-06b242db741f service nova] Acquired lock "refresh_cache-b5877d39-2f94-4a01-9db1-89a3e612095c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.459904] env[62753]: DEBUG nova.network.neutron [req-a212e5b0-7b62-4a99-8f14-11ab8ca9dac3 req-2582ef08-e74e-47de-af91-06b242db741f service nova] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Refreshing network info cache for port 96621c20-3355-4040-ad41-85fbfded9580 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1014.577913] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332641, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.719458] env[62753]: DEBUG oslo_vmware.api [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332642, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.741370] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332640, 'name': CopyVirtualDisk_Task} progress is 38%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.796605] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5211fa2a-088d-909b-d9ac-101657136c26, 'name': SearchDatastore_Task, 'duration_secs': 0.024519} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1014.796995] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1014.797378] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1014.797675] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1014.797872] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1014.798157] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1014.798725] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ff6361c4-52fc-4c73-a322-b739c9a612b8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.809935] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1014.810169] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1014.811060] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2ab14135-5e8b-4424-8d47-d7ab0b2cdf9c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.817239] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1014.817239] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52298d42-9900-6f2b-397f-b94a30ddfe11" [ 1014.817239] env[62753]: _type = "Task" [ 1014.817239] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.828125] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52298d42-9900-6f2b-397f-b94a30ddfe11, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.836142] env[62753]: DEBUG nova.objects.base [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Object Instance<2b90824a-5d49-4ec9-bcf6-4f868ca876cf> lazy-loaded attributes: info_cache,migration_context {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1014.837231] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ab9c7d7-5a15-4d87-98b4-46ed22a76ddd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.859456] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2884f534-6f60-469d-b99b-2fae874db55d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1014.865767] env[62753]: DEBUG oslo_vmware.api [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 1014.865767] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5283353f-b0d7-9467-ca8b-db2463309b96" [ 1014.865767] env[62753]: _type = "Task" [ 1014.865767] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1014.875360] env[62753]: DEBUG oslo_vmware.api [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5283353f-b0d7-9467-ca8b-db2463309b96, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1014.959190] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1015.078338] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332641, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.327354} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.078635] env[62753]: INFO nova.virt.vmwareapi.ds_util [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 9d7066e1-3721-42d2-8e80-91eacbbd8b80/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk. [ 1015.079456] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0ef383-16bb-4dcc-97d6-1d5e20e0a3cd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.107920] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Reconfiguring VM instance instance-0000005b to attach disk [datastore2] 9d7066e1-3721-42d2-8e80-91eacbbd8b80/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1015.110655] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a2f59197-7cbc-4883-b116-4ee7b5b5ee22 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.129620] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1015.129620] env[62753]: value = "task-1332644" [ 1015.129620] env[62753]: _type = "Task" [ 1015.129620] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.139581] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332644, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.217996] env[62753]: DEBUG oslo_vmware.api [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332642, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.244663] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332640, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.281926] env[62753]: DEBUG nova.network.neutron [req-a212e5b0-7b62-4a99-8f14-11ab8ca9dac3 req-2582ef08-e74e-47de-af91-06b242db741f service nova] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Updated VIF entry in instance network info cache for port 96621c20-3355-4040-ad41-85fbfded9580. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1015.282326] env[62753]: DEBUG nova.network.neutron [req-a212e5b0-7b62-4a99-8f14-11ab8ca9dac3 req-2582ef08-e74e-47de-af91-06b242db741f service nova] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Updating instance_info_cache with network_info: [{"id": "96621c20-3355-4040-ad41-85fbfded9580", "address": "fa:16:3e:a7:09:43", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap96621c20-33", "ovs_interfaceid": "96621c20-3355-4040-ad41-85fbfded9580", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1015.327725] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52298d42-9900-6f2b-397f-b94a30ddfe11, 'name': SearchDatastore_Task, 'duration_secs': 0.022904} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.328886] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-89e95ee6-1086-4848-9629-cae147f21967 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.334688] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1015.334688] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dfe762-65b3-3071-70a2-7ff4f4319b0a" [ 1015.334688] env[62753]: _type = "Task" [ 1015.334688] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.342528] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dfe762-65b3-3071-70a2-7ff4f4319b0a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.375762] env[62753]: DEBUG oslo_vmware.api [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5283353f-b0d7-9467-ca8b-db2463309b96, 'name': SearchDatastore_Task, 'duration_secs': 0.046597} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.376077] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1015.465836] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1015.466049] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.868s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1015.466371] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 3.166s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1015.640300] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332644, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.717301] env[62753]: DEBUG oslo_vmware.api [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332642, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.742634] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332640, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.785629] env[62753]: DEBUG oslo_concurrency.lockutils [req-a212e5b0-7b62-4a99-8f14-11ab8ca9dac3 req-2582ef08-e74e-47de-af91-06b242db741f service nova] Releasing lock "refresh_cache-b5877d39-2f94-4a01-9db1-89a3e612095c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.846013] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dfe762-65b3-3071-70a2-7ff4f4319b0a, 'name': SearchDatastore_Task, 'duration_secs': 0.023236} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1015.846083] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1015.846555] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] b5877d39-2f94-4a01-9db1-89a3e612095c/b5877d39-2f94-4a01-9db1-89a3e612095c.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1015.846717] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6324f1e1-4b64-45a2-8166-c9d3ec1d05f9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1015.854498] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1015.854498] env[62753]: value = "task-1332645" [ 1015.854498] env[62753]: _type = "Task" [ 1015.854498] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1015.863361] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332645, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1015.971790] env[62753]: INFO nova.compute.claims [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1016.145955] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332644, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.216807] env[62753]: DEBUG oslo_vmware.api [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332642, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.243279] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332640, 'name': CopyVirtualDisk_Task, 'duration_secs': 3.003922} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.243562] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614/3b2f7019-a7ab-4e0e-9ea9-07c8a7334614.vmdk to [datastore2] 4d20e083-2959-453a-8875-47955bc02613/4d20e083-2959-453a-8875-47955bc02613.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1016.244462] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba66d405-eb65-4e02-9425-1c271cbf3fd1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.267642] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Reconfiguring VM instance instance-00000041 to attach disk [datastore2] 4d20e083-2959-453a-8875-47955bc02613/4d20e083-2959-453a-8875-47955bc02613.vmdk or device None with type streamOptimized {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1016.268334] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aba33839-faf5-41e1-b9c0-5b67865df612 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.289259] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 1016.289259] env[62753]: value = "task-1332646" [ 1016.289259] env[62753]: _type = "Task" [ 1016.289259] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.297096] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332646, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.364768] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332645, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.479069] env[62753]: INFO nova.compute.resource_tracker [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating resource usage from migration 1c84581b-3dc0-4087-a0c2-fa39eb566788 [ 1016.651971] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332644, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.704281] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7a9a55c-fbfd-4e41-afb1-0f49004b70f7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.716874] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b7c0d57a-43e1-4aaf-90c3-77af25ed3a52 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.724623] env[62753]: DEBUG oslo_vmware.api [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332642, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.758867] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23b0ef49-b29c-4943-874f-56f43429e45f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.767587] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd3f0ed8-ad12-4c73-b8b6-309b41658296 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.783390] env[62753]: DEBUG nova.compute.provider_tree [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1016.799305] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332646, 'name': ReconfigVM_Task, 'duration_secs': 0.282054} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.800280] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Reconfigured VM instance instance-00000041 to attach disk [datastore2] 4d20e083-2959-453a-8875-47955bc02613/4d20e083-2959-453a-8875-47955bc02613.vmdk or device None with type streamOptimized {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1016.801027] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-e52b78b3-e48a-4f68-9301-d79a34756cc7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.810584] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 1016.810584] env[62753]: value = "task-1332647" [ 1016.810584] env[62753]: _type = "Task" [ 1016.810584] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.818979] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332647, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1016.867743] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332645, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.007066} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1016.868411] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] b5877d39-2f94-4a01-9db1-89a3e612095c/b5877d39-2f94-4a01-9db1-89a3e612095c.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1016.868687] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1016.868974] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b4487b16-2ad8-454d-8b70-9c73374abe8f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1016.875362] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1016.875362] env[62753]: value = "task-1332648" [ 1016.875362] env[62753]: _type = "Task" [ 1016.875362] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1016.883561] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332648, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.143451] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332644, 'name': ReconfigVM_Task, 'duration_secs': 1.619114} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.143730] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Reconfigured VM instance instance-0000005b to attach disk [datastore2] 9d7066e1-3721-42d2-8e80-91eacbbd8b80/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1017.144592] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2415f44b-1589-4db2-9c00-5b94faa67e13 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.169834] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a1561774-aedf-4b14-9387-f951a67ad101 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.186298] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1017.186298] env[62753]: value = "task-1332649" [ 1017.186298] env[62753]: _type = "Task" [ 1017.186298] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.194561] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332649, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.217096] env[62753]: DEBUG oslo_vmware.api [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332642, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.287969] env[62753]: DEBUG nova.scheduler.client.report [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1017.321532] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332647, 'name': Rename_Task, 'duration_secs': 0.134554} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.322053] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1017.322364] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-579d737f-3cfc-43a3-875c-d6f08bcef5f3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.330285] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 1017.330285] env[62753]: value = "task-1332650" [ 1017.330285] env[62753]: _type = "Task" [ 1017.330285] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.336747] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332650, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.384593] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332648, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.106839} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.384860] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1017.385674] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-04ac3309-4044-4069-a2a2-cd859688275d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.407464] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Reconfiguring VM instance instance-0000005c to attach disk [datastore1] b5877d39-2f94-4a01-9db1-89a3e612095c/b5877d39-2f94-4a01-9db1-89a3e612095c.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1017.407780] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1623fdd6-7976-4b2d-9cad-48d6a9ec67e6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.427437] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1017.427437] env[62753]: value = "task-1332651" [ 1017.427437] env[62753]: _type = "Task" [ 1017.427437] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.435996] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332651, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.696569] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332649, 'name': ReconfigVM_Task, 'duration_secs': 0.39941} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.696922] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1017.697233] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d342abae-28eb-4152-a16a-7043336af8f3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.703824] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1017.703824] env[62753]: value = "task-1332652" [ 1017.703824] env[62753]: _type = "Task" [ 1017.703824] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.717650] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332652, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.722676] env[62753]: DEBUG oslo_vmware.api [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332642, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.792915] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.326s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1017.793307] env[62753]: INFO nova.compute.manager [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Migrating [ 1017.801253] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 4.273s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1017.801558] env[62753]: DEBUG nova.objects.instance [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lazy-loading 'resources' on Instance uuid 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1017.840252] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332650, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1017.937595] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332651, 'name': ReconfigVM_Task, 'duration_secs': 0.384082} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1017.937834] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Reconfigured VM instance instance-0000005c to attach disk [datastore1] b5877d39-2f94-4a01-9db1-89a3e612095c/b5877d39-2f94-4a01-9db1-89a3e612095c.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1017.938731] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-91c68c5e-ad1e-4d73-859f-39e0c531acc7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1017.946361] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1017.946361] env[62753]: value = "task-1332653" [ 1017.946361] env[62753]: _type = "Task" [ 1017.946361] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1017.956523] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332653, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.214351] env[62753]: DEBUG oslo_vmware.api [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332652, 'name': PowerOnVM_Task, 'duration_secs': 0.405781} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.217444] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1018.221015] env[62753]: DEBUG nova.compute.manager [None req-b59e3bd6-f319-4147-ad53-fa2c086ead83 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1018.221015] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e72455c0-a556-4c22-b5c6-364dd2f2380b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.227938] env[62753]: DEBUG oslo_vmware.api [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332642, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.313254] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.313570] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.316721] env[62753]: DEBUG nova.network.neutron [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1018.342063] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332650, 'name': PowerOnVM_Task} progress is 33%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.460253] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332653, 'name': Rename_Task, 'duration_secs': 0.128676} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.463705] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1018.464186] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1018.464342] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Starting heal instance info cache {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1018.468024] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d7bf68df-c789-4571-b0f0-65c6110cff37 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.472610] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1018.472610] env[62753]: value = "task-1332654" [ 1018.472610] env[62753]: _type = "Task" [ 1018.472610] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1018.481466] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332654, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.516929] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99e1f9e9-fc49-4b62-9ebd-0c28d442a4f5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.524140] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3d212e4-a812-4124-8479-0a38af204d45 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.558730] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a39acb3-966f-4f9b-bec7-6722e33ec462 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.566875] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed0fb98b-9a8b-4c3b-97d8-1bed77b58a24 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1018.579329] env[62753]: DEBUG nova.compute.provider_tree [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1018.725763] env[62753]: DEBUG oslo_vmware.api [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332642, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.839352] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332650, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1018.969507] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1018.969667] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquired lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1018.969818] env[62753]: DEBUG nova.network.neutron [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 4d20e083-2959-453a-8875-47955bc02613] Forcefully refreshing network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1018.984096] env[62753]: DEBUG oslo_vmware.api [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332654, 'name': PowerOnVM_Task, 'duration_secs': 0.507178} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1018.984096] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1018.984096] env[62753]: INFO nova.compute.manager [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Took 7.95 seconds to spawn the instance on the hypervisor. [ 1018.984336] env[62753]: DEBUG nova.compute.manager [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1018.984959] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4ce7da62-876a-4fa1-bedf-7508bc70290b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.085022] env[62753]: DEBUG nova.scheduler.client.report [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1019.118843] env[62753]: DEBUG nova.network.neutron [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating instance_info_cache with network_info: [{"id": "81591716-fb4a-427f-b3e9-a8b265fe1ce1", "address": "fa:16:3e:3e:ef:2c", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81591716-fb", "ovs_interfaceid": "81591716-fb4a-427f-b3e9-a8b265fe1ce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1019.223693] env[62753]: DEBUG oslo_vmware.api [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332642, 'name': ReconfigVM_Task} progress is 18%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1019.339950] env[62753]: DEBUG oslo_vmware.api [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332650, 'name': PowerOnVM_Task, 'duration_secs': 1.557779} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.340316] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1019.443494] env[62753]: DEBUG nova.compute.manager [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1019.444550] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-529eed96-4e57-441f-94d4-eaabbf70ee8f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1019.512271] env[62753]: INFO nova.compute.manager [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Took 12.81 seconds to build instance. [ 1019.590835] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.790s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.593206] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.955s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.595708] env[62753]: INFO nova.compute.claims [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1019.621500] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.630505] env[62753]: INFO nova.scheduler.client.report [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Deleted allocations for instance 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5 [ 1019.694465] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "ff50c5f4-ab27-4f17-948e-80e2dcc623eb" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.695092] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "ff50c5f4-ab27-4f17-948e-80e2dcc623eb" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.735059] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "d7ba8be1-befb-4a0b-9c35-d86e7d586203" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.735526] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "d7ba8be1-befb-4a0b-9c35-d86e7d586203" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.737341] env[62753]: DEBUG oslo_vmware.api [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332642, 'name': ReconfigVM_Task, 'duration_secs': 5.840179} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1019.738247] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1019.738575] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Reconfigured VM to detach interface {{(pid=62753) detach_interface /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1978}} [ 1019.741720] env[62753]: INFO nova.compute.manager [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Unrescuing [ 1019.742167] env[62753]: DEBUG oslo_concurrency.lockutils [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "refresh_cache-9d7066e1-3721-42d2-8e80-91eacbbd8b80" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1019.742337] env[62753]: DEBUG oslo_concurrency.lockutils [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquired lock "refresh_cache-9d7066e1-3721-42d2-8e80-91eacbbd8b80" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1019.742677] env[62753]: DEBUG nova.network.neutron [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1019.765579] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "961ced74-10af-4443-857e-545cd890674e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1019.765813] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "961ced74-10af-4443-857e-545cd890674e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1019.766425] env[62753]: WARNING oslo_messaging._drivers.amqpdriver [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Number of call queues is 11, greater than warning threshold: 10. There could be a leak. Increasing threshold to: 20 [ 1019.964528] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b60befbc-103b-46c8-bc7e-0e721af48195 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "4d20e083-2959-453a-8875-47955bc02613" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 29.936s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1019.992866] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-260f318f-2ba8-4229-86dd-700e93a75c40 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.000824] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-32473fa9-cf66-4d7a-92a6-d84fba791325 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Suspending the VM {{(pid=62753) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1020.001115] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-f6f00188-925a-43c2-9b6c-6fe9eca7ea22 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.007720] env[62753]: DEBUG oslo_vmware.api [None req-32473fa9-cf66-4d7a-92a6-d84fba791325 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1020.007720] env[62753]: value = "task-1332655" [ 1020.007720] env[62753]: _type = "Task" [ 1020.007720] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1020.019530] env[62753]: DEBUG oslo_concurrency.lockutils [None req-14081901-7a08-4804-9908-fc0b637c23af tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "b5877d39-2f94-4a01-9db1-89a3e612095c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.330s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.019786] env[62753]: DEBUG oslo_vmware.api [None req-32473fa9-cf66-4d7a-92a6-d84fba791325 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332655, 'name': SuspendVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.137605] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bc9f7a4e-b277-47b6-a878-b4ec2fd958a5 tempest-AttachVolumeShelveTestJSON-577961976 tempest-AttachVolumeShelveTestJSON-577961976-project-member] Lock "4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 9.798s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1020.199368] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1020.241520] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1020.267865] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1020.345131] env[62753]: DEBUG nova.network.neutron [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 4d20e083-2959-453a-8875-47955bc02613] Updating instance_info_cache with network_info: [{"id": "9c52b1e4-82d2-4449-91f3-a624596bdc79", "address": "fa:16:3e:8e:1b:de", "network": {"id": "12446a5e-7329-4f8f-aa28-a8b9d78e7573", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1766107072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "371a5e8b16be4cffbdeb2139757684fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7c7f16b-a2e3-4d1f-9b7e-a44a9ebb589c", "external-id": "nsx-vlan-transportzone-772", "segmentation_id": 772, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c52b1e4-82", "ovs_interfaceid": "9c52b1e4-82d2-4449-91f3-a624596bdc79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.496315] env[62753]: DEBUG nova.network.neutron [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Updating instance_info_cache with network_info: [{"id": "7f800ba5-2f01-46b7-b84a-88f900822bf2", "address": "fa:16:3e:03:90:c5", "network": {"id": "22a90636-182f-4944-893f-329f446b07c4", "bridge": "br-int", "label": "tempest-ServerRescueTestJSON-1513076888-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": false}}], "meta": {"injected": false, "tenant_id": "37368e15fa5a4a0a97ac1c22fcf0bb80", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "8e7f6f41-f4eb-4832-a390-730fca1cf717", "external-id": "nsx-vlan-transportzone-724", "segmentation_id": 724, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7f800ba5-2f", "ovs_interfaceid": "7f800ba5-2f01-46b7-b84a-88f900822bf2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1020.518111] env[62753]: DEBUG oslo_vmware.api [None req-32473fa9-cf66-4d7a-92a6-d84fba791325 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332655, 'name': SuspendVM_Task} progress is 62%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1020.725415] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.767566] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.792844] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1020.848064] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48d5bdf3-c38a-44a1-9c18-cf8f1f4fe86e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.852795] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Releasing lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.853299] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 4d20e083-2959-453a-8875-47955bc02613] Updated the network info_cache for instance {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1020.859665] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f527af53-1c03-4c7a-ae70-ffc7060fb440 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.897379] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bef930a1-ab25-4370-ba9c-ed6af6743420 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.905897] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db69f96e-0741-4564-b42d-482b355caa41 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1020.920808] env[62753]: DEBUG nova.compute.provider_tree [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1020.998055] env[62753]: DEBUG oslo_concurrency.lockutils [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Releasing lock "refresh_cache-9d7066e1-3721-42d2-8e80-91eacbbd8b80" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1020.998822] env[62753]: DEBUG nova.objects.instance [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lazy-loading 'flavor' on Instance uuid 9d7066e1-3721-42d2-8e80-91eacbbd8b80 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1021.018502] env[62753]: DEBUG oslo_vmware.api [None req-32473fa9-cf66-4d7a-92a6-d84fba791325 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332655, 'name': SuspendVM_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.141127] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6271c572-b1d2-4b12-87cc-3ac4415c2d3a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.160243] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating instance 'c73b1ae0-dc98-47f7-babf-e96169384785' progress to 0 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1021.200295] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-783b6fcb-738b-49ca-bd21-7be95a486c6f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.208667] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4297abfb-7647-4dd6-be46-81762adf4b88 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Suspending the VM {{(pid=62753) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1021.208920] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-864da7db-0df6-49aa-8d13-8763b7f207c2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.215185] env[62753]: DEBUG oslo_vmware.api [None req-4297abfb-7647-4dd6-be46-81762adf4b88 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 1021.215185] env[62753]: value = "task-1332656" [ 1021.215185] env[62753]: _type = "Task" [ 1021.215185] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.222707] env[62753]: DEBUG oslo_vmware.api [None req-4297abfb-7647-4dd6-be46-81762adf4b88 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332656, 'name': SuspendVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.260184] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1021.260532] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquired lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1021.260808] env[62753]: DEBUG nova.network.neutron [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1021.425038] env[62753]: DEBUG nova.scheduler.client.report [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1021.504653] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb30b82e-0236-4af4-aa0b-dc3bc317b4cf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.528569] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1021.529414] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-690d8d58-f36f-4da0-9f10-1a1bd00cf215 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.534103] env[62753]: DEBUG oslo_vmware.api [None req-32473fa9-cf66-4d7a-92a6-d84fba791325 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332655, 'name': SuspendVM_Task, 'duration_secs': 1.029257} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1021.534719] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-32473fa9-cf66-4d7a-92a6-d84fba791325 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Suspended the VM {{(pid=62753) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1021.534912] env[62753]: DEBUG nova.compute.manager [None req-32473fa9-cf66-4d7a-92a6-d84fba791325 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1021.535717] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c76c3a4-355a-4511-baa6-a241bb6b8ba0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.539711] env[62753]: DEBUG oslo_vmware.api [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1021.539711] env[62753]: value = "task-1332657" [ 1021.539711] env[62753]: _type = "Task" [ 1021.539711] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.552100] env[62753]: DEBUG oslo_vmware.api [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332657, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.666510] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1021.666854] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-5783c0b5-3111-4262-ac3b-3336d468199f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.670432] env[62753]: DEBUG oslo_concurrency.lockutils [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.671180] env[62753]: DEBUG oslo_concurrency.lockutils [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.671404] env[62753]: DEBUG oslo_concurrency.lockutils [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1021.671604] env[62753]: DEBUG oslo_concurrency.lockutils [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1021.671778] env[62753]: DEBUG oslo_concurrency.lockutils [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.674707] env[62753]: DEBUG oslo_vmware.api [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1021.674707] env[62753]: value = "task-1332658" [ 1021.674707] env[62753]: _type = "Task" [ 1021.674707] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.675189] env[62753]: INFO nova.compute.manager [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Terminating instance [ 1021.679891] env[62753]: DEBUG nova.compute.manager [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1021.680103] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1021.680850] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c317b03-be99-4368-89af-334afcc6f202 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.691199] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] VM already powered off {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1021.691406] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating instance 'c73b1ae0-dc98-47f7-babf-e96169384785' progress to 17 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1021.694655] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1021.695097] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-01df7edd-a08b-49ec-a36f-ceaf519c0e28 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1021.701107] env[62753]: DEBUG oslo_vmware.api [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 1021.701107] env[62753]: value = "task-1332659" [ 1021.701107] env[62753]: _type = "Task" [ 1021.701107] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1021.709390] env[62753]: DEBUG oslo_vmware.api [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332659, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.725768] env[62753]: DEBUG oslo_vmware.api [None req-4297abfb-7647-4dd6-be46-81762adf4b88 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332656, 'name': SuspendVM_Task} progress is 41%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1021.930273] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.337s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1021.930962] env[62753]: DEBUG nova.compute.manager [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1021.935978] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 6.559s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.019608] env[62753]: INFO nova.network.neutron [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Port 2498c430-c268-4e1e-b331-0a3b05eeae0a from network info_cache is no longer associated with instance in Neutron. Removing from network info_cache. [ 1022.019991] env[62753]: DEBUG nova.network.neutron [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Updating instance_info_cache with network_info: [{"id": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "address": "fa:16:3e:6b:34:1e", "network": {"id": "8a9463a1-a861-4ade-8483-640392be0a89", "bridge": "br-int", "label": "tempest-AttachInterfacesTestJSON-1534757794-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.150", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b8e81660b30408c998e412f5fa81469", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "d48f0ef6-34e5-44d4-8baf-4470ed96ce73", "external-id": "nsx-vlan-transportzone-316", "segmentation_id": 316, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0cd05bfa-82", "ovs_interfaceid": "0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1022.051759] env[62753]: DEBUG oslo_vmware.api [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332657, 'name': PowerOffVM_Task, 'duration_secs': 0.288353} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.052097] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1022.057761] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Reconfiguring VM instance instance-0000005b to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1022.060129] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1066ceae-6bf0-40cb-bc96-e32a9bb16fd6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.078507] env[62753]: DEBUG oslo_vmware.api [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1022.078507] env[62753]: value = "task-1332660" [ 1022.078507] env[62753]: _type = "Task" [ 1022.078507] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.086837] env[62753]: DEBUG oslo_vmware.api [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332660, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.198168] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1022.198513] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1022.198726] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1022.198953] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1022.199156] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1022.199345] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1022.199590] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1022.199791] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1022.199980] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1022.200168] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1022.200354] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1022.205424] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bdee70c2-b447-40e2-bbfb-84dc731d9671 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.223553] env[62753]: DEBUG oslo_vmware.api [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1022.223553] env[62753]: value = "task-1332661" [ 1022.223553] env[62753]: _type = "Task" [ 1022.223553] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.232206] env[62753]: DEBUG oslo_vmware.api [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332659, 'name': PowerOffVM_Task, 'duration_secs': 0.225869} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.232480] env[62753]: DEBUG oslo_vmware.api [None req-4297abfb-7647-4dd6-be46-81762adf4b88 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332656, 'name': SuspendVM_Task, 'duration_secs': 0.84359} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.233145] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1022.233339] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1022.233619] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4297abfb-7647-4dd6-be46-81762adf4b88 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Suspended the VM {{(pid=62753) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1022.233847] env[62753]: DEBUG nova.compute.manager [None req-4297abfb-7647-4dd6-be46-81762adf4b88 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1022.234200] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-93a693dc-6744-4524-a937-73ac79ca00cc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.236101] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-af165146-5147-4ecd-bbab-ef5e95fcf6a5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.241699] env[62753]: DEBUG oslo_vmware.api [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332661, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.302171] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1022.302433] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1022.302627] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Deleting the datastore file [datastore1] 53a76eb0-8b0c-4e47-8387-75df4c0b191f {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1022.302900] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-dfd5f9f0-3819-459a-be27-829f282930cf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.309324] env[62753]: DEBUG oslo_vmware.api [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 1022.309324] env[62753]: value = "task-1332663" [ 1022.309324] env[62753]: _type = "Task" [ 1022.309324] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.317142] env[62753]: DEBUG oslo_vmware.api [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332663, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.439024] env[62753]: DEBUG nova.compute.utils [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1022.439024] env[62753]: DEBUG nova.compute.manager [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1022.439024] env[62753]: DEBUG nova.network.neutron [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1022.463340] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "b5877d39-2f94-4a01-9db1-89a3e612095c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.463340] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "b5877d39-2f94-4a01-9db1-89a3e612095c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.463340] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "b5877d39-2f94-4a01-9db1-89a3e612095c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1022.463340] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "b5877d39-2f94-4a01-9db1-89a3e612095c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1022.463340] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "b5877d39-2f94-4a01-9db1-89a3e612095c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1022.464337] env[62753]: INFO nova.compute.manager [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Terminating instance [ 1022.471281] env[62753]: DEBUG nova.compute.manager [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1022.471281] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1022.472161] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-203c3137-8064-46ac-8deb-32eecc59e775 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.480541] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1022.481028] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-356776c8-3a31-4e4b-a42a-7e8bd707773d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.515303] env[62753]: DEBUG nova.policy [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'f4dbc63095b0464fa0d2de86e3cf170e', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '5636da907ab343f9b42dbbd903d32283', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1022.524513] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Releasing lock "refresh_cache-53a76eb0-8b0c-4e47-8387-75df4c0b191f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1022.547824] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1022.547824] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1022.547824] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Deleting the datastore file [datastore1] b5877d39-2f94-4a01-9db1-89a3e612095c {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1022.547824] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ad51314a-5616-4079-a815-1441842ed896 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.562204] env[62753]: DEBUG oslo_vmware.api [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1022.562204] env[62753]: value = "task-1332665" [ 1022.562204] env[62753]: _type = "Task" [ 1022.562204] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.568179] env[62753]: DEBUG oslo_vmware.api [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332665, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.591216] env[62753]: DEBUG oslo_vmware.api [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332660, 'name': ReconfigVM_Task, 'duration_secs': 0.212937} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.591546] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Reconfigured VM instance instance-0000005b to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1022.591936] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1022.592401] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0e161818-6fec-4700-9981-4777aba8a66c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.598950] env[62753]: DEBUG oslo_vmware.api [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1022.598950] env[62753]: value = "task-1332666" [ 1022.598950] env[62753]: _type = "Task" [ 1022.598950] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1022.612154] env[62753]: DEBUG oslo_vmware.api [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332666, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1022.665902] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f25f5656-5a93-49e6-a58b-30893894a475 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.674566] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ed064cce-cbfd-4cc5-82b7-0d21c7b4d3fd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.713100] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b6cfe6ef-b448-48d5-8154-1138f9878c67 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.720576] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-668fd873-1a59-4ee9-baed-96e78c4f39c0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1022.741385] env[62753]: DEBUG nova.compute.provider_tree [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1022.745831] env[62753]: DEBUG oslo_vmware.api [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332661, 'name': ReconfigVM_Task, 'duration_secs': 0.141295} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.746420] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating instance 'c73b1ae0-dc98-47f7-babf-e96169384785' progress to 33 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1022.823275] env[62753]: DEBUG oslo_vmware.api [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332663, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.175846} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1022.823275] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1022.823275] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1022.823275] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1022.823275] env[62753]: INFO nova.compute.manager [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1022.823275] env[62753]: DEBUG oslo.service.loopingcall [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1022.823275] env[62753]: DEBUG nova.compute.manager [-] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1022.823275] env[62753]: DEBUG nova.network.neutron [-] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1022.942594] env[62753]: DEBUG nova.compute.manager [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1023.026855] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aaf20174-5852-4dc7-8698-2445abad4950 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "interface-53a76eb0-8b0c-4e47-8387-75df4c0b191f-2498c430-c268-4e1e-b331-0a3b05eeae0a" "released" by "nova.compute.manager.ComputeManager.detach_interface..do_detach_interface" :: held 9.900s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1023.070473] env[62753]: DEBUG oslo_vmware.api [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332665, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.181805} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.070813] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1023.071014] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1023.071210] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1023.071436] env[62753]: INFO nova.compute.manager [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Took 0.60 seconds to destroy the instance on the hypervisor. [ 1023.071724] env[62753]: DEBUG oslo.service.loopingcall [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1023.071922] env[62753]: DEBUG nova.compute.manager [-] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1023.072027] env[62753]: DEBUG nova.network.neutron [-] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1023.109101] env[62753]: DEBUG oslo_vmware.api [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332666, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.248023] env[62753]: DEBUG nova.scheduler.client.report [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1023.255599] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1023.255599] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1023.255599] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1023.255599] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1023.255920] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1023.255920] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1023.256059] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1023.256201] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1023.256570] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1023.256570] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1023.257102] env[62753]: DEBUG nova.virt.hardware [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1023.262716] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Reconfiguring VM instance instance-00000048 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1023.266830] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b4659ff9-dd9f-49d0-bb45-9262ef9b14fc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.284261] env[62753]: DEBUG oslo_vmware.api [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1023.284261] env[62753]: value = "task-1332667" [ 1023.284261] env[62753]: _type = "Task" [ 1023.284261] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.294018] env[62753]: DEBUG oslo_vmware.api [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332667, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.339570] env[62753]: DEBUG nova.network.neutron [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Successfully created port: b3e995ee-b33d-47b0-8302-8171441ddfd6 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1023.611107] env[62753]: DEBUG oslo_vmware.api [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332666, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.797739] env[62753]: DEBUG oslo_vmware.api [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332667, 'name': ReconfigVM_Task, 'duration_secs': 0.249506} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1023.798740] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Reconfigured VM instance instance-00000048 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1023.799562] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-374f7d2a-5c44-48a9-a063-6d66c9aea6f0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.822694] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Reconfiguring VM instance instance-00000048 to attach disk [datastore1] c73b1ae0-dc98-47f7-babf-e96169384785/c73b1ae0-dc98-47f7-babf-e96169384785.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1023.823309] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ac3b13de-dbd9-4f04-a163-f0163cfe47b3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.848053] env[62753]: INFO nova.compute.manager [None req-b72fb039-f978-41b7-9e19-1887bc5107da tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Resuming [ 1023.848673] env[62753]: DEBUG nova.objects.instance [None req-b72fb039-f978-41b7-9e19-1887bc5107da tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lazy-loading 'flavor' on Instance uuid 4d20e083-2959-453a-8875-47955bc02613 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1023.853034] env[62753]: DEBUG oslo_vmware.api [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1023.853034] env[62753]: value = "task-1332668" [ 1023.853034] env[62753]: _type = "Task" [ 1023.853034] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1023.864524] env[62753]: DEBUG oslo_vmware.api [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332668, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1023.869252] env[62753]: DEBUG nova.compute.manager [req-0bc0e7f1-49c0-466b-ab33-9ba4a454be0f req-cee77795-a030-4bd1-a967-ba621a7d34d8 service nova] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Received event network-vif-deleted-96621c20-3355-4040-ad41-85fbfded9580 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1023.869756] env[62753]: INFO nova.compute.manager [req-0bc0e7f1-49c0-466b-ab33-9ba4a454be0f req-cee77795-a030-4bd1-a967-ba621a7d34d8 service nova] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Neutron deleted interface 96621c20-3355-4040-ad41-85fbfded9580; detaching it from the instance and deleting it from the info cache [ 1023.869995] env[62753]: DEBUG nova.network.neutron [req-0bc0e7f1-49c0-466b-ab33-9ba4a454be0f req-cee77795-a030-4bd1-a967-ba621a7d34d8 service nova] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1023.953299] env[62753]: DEBUG nova.compute.manager [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1023.986242] env[62753]: DEBUG nova.virt.hardware [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1023.986587] env[62753]: DEBUG nova.virt.hardware [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1023.986730] env[62753]: DEBUG nova.virt.hardware [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1023.986953] env[62753]: DEBUG nova.virt.hardware [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1023.987166] env[62753]: DEBUG nova.virt.hardware [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1023.987376] env[62753]: DEBUG nova.virt.hardware [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1023.987640] env[62753]: DEBUG nova.virt.hardware [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1023.987833] env[62753]: DEBUG nova.virt.hardware [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1023.988064] env[62753]: DEBUG nova.virt.hardware [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1023.988249] env[62753]: DEBUG nova.virt.hardware [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1023.988470] env[62753]: DEBUG nova.virt.hardware [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1023.989436] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74b3333e-beab-4e75-8a71-02e3a4c0332f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1023.998088] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd17aa67-722b-4f43-8d74-f43b48e16ffb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.112248] env[62753]: DEBUG oslo_vmware.api [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332666, 'name': PowerOnVM_Task, 'duration_secs': 1.0992} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.112571] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1024.115373] env[62753]: DEBUG nova.compute.manager [None req-87cb29a6-6295-4f3a-b6fc-5c5b0593a23a tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1024.115373] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99f84b61-1cfb-476b-bfe8-b563f9524c67 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.272194] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.337s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1024.275188] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.550s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1024.276663] env[62753]: INFO nova.compute.claims [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1024.342715] env[62753]: DEBUG nova.network.neutron [-] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.367980] env[62753]: DEBUG oslo_vmware.api [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332668, 'name': ReconfigVM_Task, 'duration_secs': 0.326237} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1024.368525] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Reconfigured VM instance instance-00000048 to attach disk [datastore1] c73b1ae0-dc98-47f7-babf-e96169384785/c73b1ae0-dc98-47f7-babf-e96169384785.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1024.368806] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating instance 'c73b1ae0-dc98-47f7-babf-e96169384785' progress to 50 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1024.373708] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-36414fda-1ec4-45ba-8005-397c5926a5a0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.385013] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5fce8e19-ba75-4bf5-8ae9-e740bf219abe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.419114] env[62753]: DEBUG nova.compute.manager [req-0bc0e7f1-49c0-466b-ab33-9ba4a454be0f req-cee77795-a030-4bd1-a967-ba621a7d34d8 service nova] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Detach interface failed, port_id=96621c20-3355-4040-ad41-85fbfded9580, reason: Instance b5877d39-2f94-4a01-9db1-89a3e612095c could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1024.750829] env[62753]: DEBUG nova.network.neutron [-] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1024.845748] env[62753]: INFO nova.scheduler.client.report [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Deleted allocation for migration 4a936932-16ef-44c3-898d-da10d29f220b [ 1024.847095] env[62753]: INFO nova.compute.manager [-] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Took 1.77 seconds to deallocate network for instance. [ 1024.864470] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b72fb039-f978-41b7-9e19-1887bc5107da tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1024.864649] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b72fb039-f978-41b7-9e19-1887bc5107da tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquired lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1024.864828] env[62753]: DEBUG nova.network.neutron [None req-b72fb039-f978-41b7-9e19-1887bc5107da tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1024.882444] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c6b931-0e4e-4cfe-952a-10edcf6b017d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.903597] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27e61f3d-7333-4783-8200-eaf38fdfa5ed {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1024.922714] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating instance 'c73b1ae0-dc98-47f7-babf-e96169384785' progress to 67 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1025.228185] env[62753]: DEBUG nova.network.neutron [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Successfully updated port: b3e995ee-b33d-47b0-8302-8171441ddfd6 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1025.253427] env[62753]: INFO nova.compute.manager [-] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Took 2.43 seconds to deallocate network for instance. [ 1025.351834] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6177e994-f061-4c00-94f5-910ee6f78b68 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 13.455s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.356705] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.467955] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-54e04bf1-9095-4108-80e6-3f25fb782790 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.477858] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1228a58-823f-4306-865e-96138dc44927 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.508856] env[62753]: DEBUG nova.network.neutron [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Port 81591716-fb4a-427f-b3e9-a8b265fe1ce1 binding to destination host cpu-1 is already ACTIVE {{(pid=62753) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1025.512712] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d954c729-88a9-40ec-8cd3-0e961c97a27c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.521027] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-475e2de1-ca18-4a21-b5aa-56d6b562d9f3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.535354] env[62753]: DEBUG nova.compute.provider_tree [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1025.607331] env[62753]: DEBUG nova.network.neutron [None req-b72fb039-f978-41b7-9e19-1887bc5107da tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Updating instance_info_cache with network_info: [{"id": "9c52b1e4-82d2-4449-91f3-a624596bdc79", "address": "fa:16:3e:8e:1b:de", "network": {"id": "12446a5e-7329-4f8f-aa28-a8b9d78e7573", "bridge": "br-int", "label": "tempest-ServersNegativeTestJSON-1766107072-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "371a5e8b16be4cffbdeb2139757684fc", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "c7c7f16b-a2e3-4d1f-9b7e-a44a9ebb589c", "external-id": "nsx-vlan-transportzone-772", "segmentation_id": 772, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap9c52b1e4-82", "ovs_interfaceid": "9c52b1e4-82d2-4449-91f3-a624596bdc79", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1025.730822] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "refresh_cache-b1e4c8f9-471d-4cca-9a12-56191767a5b1" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1025.731060] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "refresh_cache-b1e4c8f9-471d-4cca-9a12-56191767a5b1" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1025.731158] env[62753]: DEBUG nova.network.neutron [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1025.743395] env[62753]: DEBUG oslo_concurrency.lockutils [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "9d7066e1-3721-42d2-8e80-91eacbbd8b80" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.743641] env[62753]: DEBUG oslo_concurrency.lockutils [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "9d7066e1-3721-42d2-8e80-91eacbbd8b80" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.743852] env[62753]: DEBUG oslo_concurrency.lockutils [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "9d7066e1-3721-42d2-8e80-91eacbbd8b80-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.744053] env[62753]: DEBUG oslo_concurrency.lockutils [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "9d7066e1-3721-42d2-8e80-91eacbbd8b80-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.744269] env[62753]: DEBUG oslo_concurrency.lockutils [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "9d7066e1-3721-42d2-8e80-91eacbbd8b80-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.746467] env[62753]: INFO nova.compute.manager [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Terminating instance [ 1025.748481] env[62753]: DEBUG nova.compute.manager [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1025.748687] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1025.749555] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adbc1bb2-9ddc-4ce8-92a1-7cfe22d79feb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.759040] env[62753]: DEBUG oslo_concurrency.lockutils [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.759351] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1025.759586] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-2a8ff857-bae4-4e34-8130-f69e19f05d0b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1025.765920] env[62753]: DEBUG oslo_vmware.api [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1025.765920] env[62753]: value = "task-1332669" [ 1025.765920] env[62753]: _type = "Task" [ 1025.765920] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1025.776462] env[62753]: DEBUG oslo_vmware.api [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332669, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1025.901962] env[62753]: DEBUG nova.compute.manager [req-c1170294-0c99-4a7b-b36c-3b590251045c req-c63b6a84-06cf-4c97-8505-4886b9ce42e7 service nova] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Received event network-vif-deleted-0cd05bfa-82e2-4f11-b1d0-0f64c7065ed9 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1025.902347] env[62753]: DEBUG nova.compute.manager [req-c1170294-0c99-4a7b-b36c-3b590251045c req-c63b6a84-06cf-4c97-8505-4886b9ce42e7 service nova] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Received event network-vif-plugged-b3e995ee-b33d-47b0-8302-8171441ddfd6 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1025.902463] env[62753]: DEBUG oslo_concurrency.lockutils [req-c1170294-0c99-4a7b-b36c-3b590251045c req-c63b6a84-06cf-4c97-8505-4886b9ce42e7 service nova] Acquiring lock "b1e4c8f9-471d-4cca-9a12-56191767a5b1-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1025.902740] env[62753]: DEBUG oslo_concurrency.lockutils [req-c1170294-0c99-4a7b-b36c-3b590251045c req-c63b6a84-06cf-4c97-8505-4886b9ce42e7 service nova] Lock "b1e4c8f9-471d-4cca-9a12-56191767a5b1-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1025.902977] env[62753]: DEBUG oslo_concurrency.lockutils [req-c1170294-0c99-4a7b-b36c-3b590251045c req-c63b6a84-06cf-4c97-8505-4886b9ce42e7 service nova] Lock "b1e4c8f9-471d-4cca-9a12-56191767a5b1-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1025.903206] env[62753]: DEBUG nova.compute.manager [req-c1170294-0c99-4a7b-b36c-3b590251045c req-c63b6a84-06cf-4c97-8505-4886b9ce42e7 service nova] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] No waiting events found dispatching network-vif-plugged-b3e995ee-b33d-47b0-8302-8171441ddfd6 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1025.903447] env[62753]: WARNING nova.compute.manager [req-c1170294-0c99-4a7b-b36c-3b590251045c req-c63b6a84-06cf-4c97-8505-4886b9ce42e7 service nova] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Received unexpected event network-vif-plugged-b3e995ee-b33d-47b0-8302-8171441ddfd6 for instance with vm_state building and task_state spawning. [ 1025.903713] env[62753]: DEBUG nova.compute.manager [req-c1170294-0c99-4a7b-b36c-3b590251045c req-c63b6a84-06cf-4c97-8505-4886b9ce42e7 service nova] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Received event network-changed-b3e995ee-b33d-47b0-8302-8171441ddfd6 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1025.903848] env[62753]: DEBUG nova.compute.manager [req-c1170294-0c99-4a7b-b36c-3b590251045c req-c63b6a84-06cf-4c97-8505-4886b9ce42e7 service nova] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Refreshing instance network info cache due to event network-changed-b3e995ee-b33d-47b0-8302-8171441ddfd6. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1025.904087] env[62753]: DEBUG oslo_concurrency.lockutils [req-c1170294-0c99-4a7b-b36c-3b590251045c req-c63b6a84-06cf-4c97-8505-4886b9ce42e7 service nova] Acquiring lock "refresh_cache-b1e4c8f9-471d-4cca-9a12-56191767a5b1" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1026.038828] env[62753]: DEBUG nova.scheduler.client.report [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1026.114224] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b72fb039-f978-41b7-9e19-1887bc5107da tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Releasing lock "refresh_cache-4d20e083-2959-453a-8875-47955bc02613" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.115362] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e51e3fa3-bc09-4fc7-8723-113c5f2a979b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.122972] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b72fb039-f978-41b7-9e19-1887bc5107da tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Resuming the VM {{(pid=62753) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1026.123104] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-3875e10e-643d-4948-aaaa-71139860f350 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.129521] env[62753]: DEBUG oslo_vmware.api [None req-b72fb039-f978-41b7-9e19-1887bc5107da tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 1026.129521] env[62753]: value = "task-1332670" [ 1026.129521] env[62753]: _type = "Task" [ 1026.129521] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.137930] env[62753]: DEBUG oslo_vmware.api [None req-b72fb039-f978-41b7-9e19-1887bc5107da tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332670, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.264508] env[62753]: DEBUG nova.network.neutron [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1026.275181] env[62753]: DEBUG oslo_vmware.api [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332669, 'name': PowerOffVM_Task, 'duration_secs': 0.230294} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.275602] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1026.275785] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1026.276057] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4f614c30-fa11-4a6b-aff1-de9d82c82350 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.366667] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1026.366922] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1026.367107] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Deleting the datastore file [datastore2] 9d7066e1-3721-42d2-8e80-91eacbbd8b80 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1026.367400] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-30fe5f37-96a8-46b7-b639-e4d57abab433 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.373739] env[62753]: DEBUG oslo_vmware.api [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1026.373739] env[62753]: value = "task-1332672" [ 1026.373739] env[62753]: _type = "Task" [ 1026.373739] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.379475] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.379475] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.379762] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.379976] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.380167] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.382261] env[62753]: INFO nova.compute.manager [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Terminating instance [ 1026.384104] env[62753]: DEBUG nova.compute.manager [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1026.384314] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1026.385182] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ada85ec-c779-4c9a-8b27-4f26d3cc89fa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.392654] env[62753]: DEBUG oslo_vmware.api [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332672, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.394884] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1026.395092] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-45c68925-7c48-47a4-837c-327a5de231f2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.400967] env[62753]: DEBUG oslo_vmware.api [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 1026.400967] env[62753]: value = "task-1332673" [ 1026.400967] env[62753]: _type = "Task" [ 1026.400967] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.408651] env[62753]: DEBUG oslo_vmware.api [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332673, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.423854] env[62753]: DEBUG nova.network.neutron [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Updating instance_info_cache with network_info: [{"id": "b3e995ee-b33d-47b0-8302-8171441ddfd6", "address": "fa:16:3e:0b:d8:71", "network": {"id": "35ed0ec0-821f-45b6-bc2d-015626bfc0a2", "bridge": "br-int", "label": "tempest-ServersTestJSON-444726172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5636da907ab343f9b42dbbd903d32283", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3e995ee-b3", "ovs_interfaceid": "b3e995ee-b33d-47b0-8302-8171441ddfd6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1026.534230] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "c73b1ae0-dc98-47f7-babf-e96169384785-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.534512] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "c73b1ae0-dc98-47f7-babf-e96169384785-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.534699] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "c73b1ae0-dc98-47f7-babf-e96169384785-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.545429] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.270s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1026.546053] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1026.549953] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.782s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.551935] env[62753]: INFO nova.compute.claims [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1026.640233] env[62753]: DEBUG oslo_vmware.api [None req-b72fb039-f978-41b7-9e19-1887bc5107da tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332670, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.888067] env[62753]: DEBUG oslo_vmware.api [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332672, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.272441} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.888253] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1026.888445] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1026.888621] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1026.888793] env[62753]: INFO nova.compute.manager [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1026.889055] env[62753]: DEBUG oslo.service.loopingcall [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.889264] env[62753]: DEBUG nova.compute.manager [-] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1026.889361] env[62753]: DEBUG nova.network.neutron [-] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1026.899942] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "0f81b85b-0550-46b9-b77c-9075a99cfaec" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1026.900322] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "0f81b85b-0550-46b9-b77c-9075a99cfaec" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1026.913978] env[62753]: DEBUG oslo_vmware.api [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332673, 'name': PowerOffVM_Task, 'duration_secs': 0.202175} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1026.914686] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1026.914885] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1026.915571] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-cd71d825-4770-42b1-987e-4a1ae05e3eb2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.926718] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "refresh_cache-b1e4c8f9-471d-4cca-9a12-56191767a5b1" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1026.927153] env[62753]: DEBUG nova.compute.manager [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Instance network_info: |[{"id": "b3e995ee-b33d-47b0-8302-8171441ddfd6", "address": "fa:16:3e:0b:d8:71", "network": {"id": "35ed0ec0-821f-45b6-bc2d-015626bfc0a2", "bridge": "br-int", "label": "tempest-ServersTestJSON-444726172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5636da907ab343f9b42dbbd903d32283", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3e995ee-b3", "ovs_interfaceid": "b3e995ee-b33d-47b0-8302-8171441ddfd6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1026.927742] env[62753]: DEBUG oslo_concurrency.lockutils [req-c1170294-0c99-4a7b-b36c-3b590251045c req-c63b6a84-06cf-4c97-8505-4886b9ce42e7 service nova] Acquired lock "refresh_cache-b1e4c8f9-471d-4cca-9a12-56191767a5b1" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1026.927938] env[62753]: DEBUG nova.network.neutron [req-c1170294-0c99-4a7b-b36c-3b590251045c req-c63b6a84-06cf-4c97-8505-4886b9ce42e7 service nova] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Refreshing network info cache for port b3e995ee-b33d-47b0-8302-8171441ddfd6 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1026.929194] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0b:d8:71', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '2624812a-9f9c-461d-8b5f-79bea90c7ad3', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b3e995ee-b33d-47b0-8302-8171441ddfd6', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1026.937517] env[62753]: DEBUG oslo.service.loopingcall [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1026.938882] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1026.939193] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cb7e215d-76bb-47b1-b8d6-96e3402887e0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.957987] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1026.957987] env[62753]: value = "task-1332675" [ 1026.957987] env[62753]: _type = "Task" [ 1026.957987] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1026.965993] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332675, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1026.979424] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1026.979655] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1026.979866] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Deleting the datastore file [datastore2] 2b90824a-5d49-4ec9-bcf6-4f868ca876cf {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1026.980150] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6abae150-129b-4152-a9a3-093dfdd444d3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1026.992223] env[62753]: DEBUG oslo_vmware.api [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 1026.992223] env[62753]: value = "task-1332676" [ 1026.992223] env[62753]: _type = "Task" [ 1026.992223] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.000915] env[62753]: DEBUG oslo_vmware.api [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332676, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.051866] env[62753]: DEBUG nova.compute.utils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1027.053265] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1027.053442] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1027.128835] env[62753]: DEBUG nova.policy [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '199d994ed55a491cb5802fa000963c85', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e511c03cdf5c4ce9bcd9e56b425e20fb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1027.139313] env[62753]: DEBUG oslo_vmware.api [None req-b72fb039-f978-41b7-9e19-1887bc5107da tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332670, 'name': PowerOnVM_Task, 'duration_secs': 0.610104} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.139583] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b72fb039-f978-41b7-9e19-1887bc5107da tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Resumed the VM {{(pid=62753) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1027.139773] env[62753]: DEBUG nova.compute.manager [None req-b72fb039-f978-41b7-9e19-1887bc5107da tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1027.140592] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8e006e26-1938-4e49-96b5-53723683599b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.402490] env[62753]: DEBUG nova.compute.manager [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1027.427654] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Successfully created port: 86406a4f-4de3-4c2d-b326-ced3cbdf893b {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1027.468256] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332675, 'name': CreateVM_Task, 'duration_secs': 0.351614} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.468441] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1027.469128] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.469305] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.469647] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1027.470717] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddbd5285-f0c0-4e31-90e4-2957168718a1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.475487] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 1027.475487] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]524d428f-09be-e13e-8039-53392bcfebc8" [ 1027.475487] env[62753]: _type = "Task" [ 1027.475487] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1027.485573] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]524d428f-09be-e13e-8039-53392bcfebc8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1027.500318] env[62753]: DEBUG oslo_vmware.api [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332676, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.181135} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.500575] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1027.500761] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1027.500956] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1027.501344] env[62753]: INFO nova.compute.manager [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1027.501604] env[62753]: DEBUG oslo.service.loopingcall [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1027.501803] env[62753]: DEBUG nova.compute.manager [-] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1027.501898] env[62753]: DEBUG nova.network.neutron [-] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1027.557243] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1027.589257] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.589460] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.589644] env[62753]: DEBUG nova.network.neutron [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1027.760317] env[62753]: DEBUG nova.network.neutron [-] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.781433] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cae0b9f0-41c4-4d14-9aeb-6bd2740af668 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.789915] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3e7b0c0b-df66-458e-96fb-cec24c9456e7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.822219] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd4166b4-99a2-4352-8dda-afa7c2761f90 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.830421] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a5b2163-82b9-41e6-b3a3-6f288cefcc5f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1027.843555] env[62753]: DEBUG nova.compute.provider_tree [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1027.921438] env[62753]: DEBUG nova.network.neutron [req-c1170294-0c99-4a7b-b36c-3b590251045c req-c63b6a84-06cf-4c97-8505-4886b9ce42e7 service nova] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Updated VIF entry in instance network info cache for port b3e995ee-b33d-47b0-8302-8171441ddfd6. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1027.921795] env[62753]: DEBUG nova.network.neutron [req-c1170294-0c99-4a7b-b36c-3b590251045c req-c63b6a84-06cf-4c97-8505-4886b9ce42e7 service nova] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Updating instance_info_cache with network_info: [{"id": "b3e995ee-b33d-47b0-8302-8171441ddfd6", "address": "fa:16:3e:0b:d8:71", "network": {"id": "35ed0ec0-821f-45b6-bc2d-015626bfc0a2", "bridge": "br-int", "label": "tempest-ServersTestJSON-444726172-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "5636da907ab343f9b42dbbd903d32283", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "2624812a-9f9c-461d-8b5f-79bea90c7ad3", "external-id": "nsx-vlan-transportzone-123", "segmentation_id": 123, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb3e995ee-b3", "ovs_interfaceid": "b3e995ee-b33d-47b0-8302-8171441ddfd6", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1027.923723] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1027.990716] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]524d428f-09be-e13e-8039-53392bcfebc8, 'name': SearchDatastore_Task, 'duration_secs': 0.025443} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1027.991043] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1027.991320] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1027.991575] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1027.991716] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1027.991897] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1027.992183] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-9125d5c9-f251-4738-8ccf-b52aa98a7fc6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.002813] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1028.003012] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1028.004132] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2f870cda-3672-4c95-a24c-1e722e9658d2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.010536] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 1028.010536] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52482f71-980c-b684-1d79-715b0376d1fb" [ 1028.010536] env[62753]: _type = "Task" [ 1028.010536] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.016844] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52482f71-980c-b684-1d79-715b0376d1fb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.048610] env[62753]: DEBUG nova.compute.manager [req-96974fbf-be1c-4766-9fb1-2b39f09c3eec req-519e0ddd-9f0d-47e5-a7d9-3f634e40c260 service nova] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Received event network-vif-deleted-7f800ba5-2f01-46b7-b84a-88f900822bf2 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1028.262415] env[62753]: INFO nova.compute.manager [-] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Took 1.37 seconds to deallocate network for instance. [ 1028.303714] env[62753]: DEBUG nova.network.neutron [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating instance_info_cache with network_info: [{"id": "81591716-fb4a-427f-b3e9-a8b265fe1ce1", "address": "fa:16:3e:3e:ef:2c", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81591716-fb", "ovs_interfaceid": "81591716-fb4a-427f-b3e9-a8b265fe1ce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.347007] env[62753]: DEBUG nova.scheduler.client.report [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1028.425934] env[62753]: DEBUG oslo_concurrency.lockutils [req-c1170294-0c99-4a7b-b36c-3b590251045c req-c63b6a84-06cf-4c97-8505-4886b9ce42e7 service nova] Releasing lock "refresh_cache-b1e4c8f9-471d-4cca-9a12-56191767a5b1" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.524776] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52482f71-980c-b684-1d79-715b0376d1fb, 'name': SearchDatastore_Task, 'duration_secs': 0.028772} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1028.525631] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3b7a5c1-0f79-4b09-8244-be0760fe6d3e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.530825] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 1028.530825] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52cf8032-3e0e-03d9-b5c6-e8b33312d2c4" [ 1028.530825] env[62753]: _type = "Task" [ 1028.530825] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1028.538338] env[62753]: DEBUG nova.network.neutron [-] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1028.539475] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52cf8032-3e0e-03d9-b5c6-e8b33312d2c4, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1028.573067] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1028.601224] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1028.601771] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1028.601771] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1028.601771] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1028.602294] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1028.602294] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1028.602587] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1028.602587] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1028.602663] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1028.602931] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1028.603050] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1028.603928] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7dc9d0b-4914-43b5-a7c3-d74827f6afb9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.611742] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a35eb3d9-0139-45ea-812e-6c90244cd76a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1028.771701] env[62753]: DEBUG oslo_concurrency.lockutils [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1028.806636] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1028.851777] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.302s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1028.852312] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1028.854788] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 8.062s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1028.856203] env[62753]: INFO nova.compute.claims [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1029.041601] env[62753]: INFO nova.compute.manager [-] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Took 1.54 seconds to deallocate network for instance. [ 1029.041949] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52cf8032-3e0e-03d9-b5c6-e8b33312d2c4, 'name': SearchDatastore_Task, 'duration_secs': 0.009174} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.043641] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1029.043901] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] b1e4c8f9-471d-4cca-9a12-56191767a5b1/b1e4c8f9-471d-4cca-9a12-56191767a5b1.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1029.046609] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fae35e47-5bb8-4bb1-9d30-a8b4821ed0e0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.055421] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 1029.055421] env[62753]: value = "task-1332677" [ 1029.055421] env[62753]: _type = "Task" [ 1029.055421] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.063813] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332677, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.285720] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Successfully updated port: 86406a4f-4de3-4c2d-b326-ced3cbdf893b {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1029.341647] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-587a1c82-b434-4a87-b66c-906b90aa50f1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.366355] env[62753]: DEBUG nova.compute.utils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1029.368072] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1029.368255] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1029.370618] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66f1590c-3acd-4cc9-a02c-901f3cc77939 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.379576] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating instance 'c73b1ae0-dc98-47f7-babf-e96169384785' progress to 83 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1029.425868] env[62753]: DEBUG nova.policy [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '199d994ed55a491cb5802fa000963c85', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e511c03cdf5c4ce9bcd9e56b425e20fb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1029.552310] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1029.566670] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332677, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.467477} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1029.566670] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] b1e4c8f9-471d-4cca-9a12-56191767a5b1/b1e4c8f9-471d-4cca-9a12-56191767a5b1.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1029.566862] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1029.567112] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-96869421-fed8-43df-8723-abfbffa92222 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1029.573056] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 1029.573056] env[62753]: value = "task-1332678" [ 1029.573056] env[62753]: _type = "Task" [ 1029.573056] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1029.580447] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332678, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1029.690690] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Successfully created port: 0c86bac1-aaf9-4933-ac5a-6488554c390c {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1029.790143] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "refresh_cache-ff50c5f4-ab27-4f17-948e-80e2dcc623eb" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1029.790143] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquired lock "refresh_cache-ff50c5f4-ab27-4f17-948e-80e2dcc623eb" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1029.790143] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1029.874332] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1029.886512] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aa72699b-4d98-49c8-9948-5095d1a07a92 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating instance 'c73b1ae0-dc98-47f7-babf-e96169384785' progress to 100 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1030.051926] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-507821d7-75a5-4253-b87e-610530e118fb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.059246] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13191848-9439-45a5-bee3-f9b1d7396933 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.094396] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e85ad3c-932d-4506-8ffd-2833f32f25cb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.097931] env[62753]: DEBUG nova.compute.manager [req-916b4084-08ab-4469-8ffd-cf22a492beb4 req-c407ef6c-e04f-4ba2-817a-a5d49b5a11a3 service nova] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Received event network-vif-deleted-c86749ca-c0b7-4c20-a32a-fed94bb47e4d {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.098150] env[62753]: DEBUG nova.compute.manager [req-916b4084-08ab-4469-8ffd-cf22a492beb4 req-c407ef6c-e04f-4ba2-817a-a5d49b5a11a3 service nova] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Received event network-vif-plugged-86406a4f-4de3-4c2d-b326-ced3cbdf893b {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.098342] env[62753]: DEBUG oslo_concurrency.lockutils [req-916b4084-08ab-4469-8ffd-cf22a492beb4 req-c407ef6c-e04f-4ba2-817a-a5d49b5a11a3 service nova] Acquiring lock "ff50c5f4-ab27-4f17-948e-80e2dcc623eb-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1030.098552] env[62753]: DEBUG oslo_concurrency.lockutils [req-916b4084-08ab-4469-8ffd-cf22a492beb4 req-c407ef6c-e04f-4ba2-817a-a5d49b5a11a3 service nova] Lock "ff50c5f4-ab27-4f17-948e-80e2dcc623eb-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1030.098720] env[62753]: DEBUG oslo_concurrency.lockutils [req-916b4084-08ab-4469-8ffd-cf22a492beb4 req-c407ef6c-e04f-4ba2-817a-a5d49b5a11a3 service nova] Lock "ff50c5f4-ab27-4f17-948e-80e2dcc623eb-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1030.098886] env[62753]: DEBUG nova.compute.manager [req-916b4084-08ab-4469-8ffd-cf22a492beb4 req-c407ef6c-e04f-4ba2-817a-a5d49b5a11a3 service nova] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] No waiting events found dispatching network-vif-plugged-86406a4f-4de3-4c2d-b326-ced3cbdf893b {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1030.099064] env[62753]: WARNING nova.compute.manager [req-916b4084-08ab-4469-8ffd-cf22a492beb4 req-c407ef6c-e04f-4ba2-817a-a5d49b5a11a3 service nova] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Received unexpected event network-vif-plugged-86406a4f-4de3-4c2d-b326-ced3cbdf893b for instance with vm_state building and task_state spawning. [ 1030.099232] env[62753]: DEBUG nova.compute.manager [req-916b4084-08ab-4469-8ffd-cf22a492beb4 req-c407ef6c-e04f-4ba2-817a-a5d49b5a11a3 service nova] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Received event network-changed-86406a4f-4de3-4c2d-b326-ced3cbdf893b {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1030.099392] env[62753]: DEBUG nova.compute.manager [req-916b4084-08ab-4469-8ffd-cf22a492beb4 req-c407ef6c-e04f-4ba2-817a-a5d49b5a11a3 service nova] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Refreshing instance network info cache due to event network-changed-86406a4f-4de3-4c2d-b326-ced3cbdf893b. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1030.099564] env[62753]: DEBUG oslo_concurrency.lockutils [req-916b4084-08ab-4469-8ffd-cf22a492beb4 req-c407ef6c-e04f-4ba2-817a-a5d49b5a11a3 service nova] Acquiring lock "refresh_cache-ff50c5f4-ab27-4f17-948e-80e2dcc623eb" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1030.104727] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332678, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.29401} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.106660] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1030.107388] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac989437-dc42-49ba-a99c-59f5674916a8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.110766] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f6ceda7-cbda-42d1-b8bc-5c5eea0d9c77 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.134506] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Reconfiguring VM instance instance-0000005d to attach disk [datastore2] b1e4c8f9-471d-4cca-9a12-56191767a5b1/b1e4c8f9-471d-4cca-9a12-56191767a5b1.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1030.141930] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-67f55239-6a80-414e-92f4-883b8c04f1e4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.155611] env[62753]: DEBUG nova.compute.provider_tree [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1030.162157] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 1030.162157] env[62753]: value = "task-1332679" [ 1030.162157] env[62753]: _type = "Task" [ 1030.162157] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.170410] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332679, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.321157] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1030.455924] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Updating instance_info_cache with network_info: [{"id": "86406a4f-4de3-4c2d-b326-ced3cbdf893b", "address": "fa:16:3e:6e:a7:eb", "network": {"id": "1fcbe5e7-7620-481b-a739-be070f31d15f", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-84347152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e511c03cdf5c4ce9bcd9e56b425e20fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86406a4f-4d", "ovs_interfaceid": "86406a4f-4de3-4c2d-b326-ced3cbdf893b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1030.658535] env[62753]: DEBUG nova.scheduler.client.report [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1030.670902] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332679, 'name': ReconfigVM_Task, 'duration_secs': 0.311663} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1030.671200] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Reconfigured VM instance instance-0000005d to attach disk [datastore2] b1e4c8f9-471d-4cca-9a12-56191767a5b1/b1e4c8f9-471d-4cca-9a12-56191767a5b1.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1030.672341] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-806671e6-b7ab-4cd1-ab0f-744c243ad816 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.678649] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 1030.678649] env[62753]: value = "task-1332680" [ 1030.678649] env[62753]: _type = "Task" [ 1030.678649] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1030.688278] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332680, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1030.888648] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1030.912345] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1030.912594] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1030.912793] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1030.913015] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1030.913196] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1030.913655] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1030.913655] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1030.913789] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1030.913976] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1030.914216] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1030.914432] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1030.915384] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f8abb6d-4ec3-4538-abde-7eb8e11d52f0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.923880] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e580262f-f541-4028-9f6b-46d76ebed28f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.958072] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Releasing lock "refresh_cache-ff50c5f4-ab27-4f17-948e-80e2dcc623eb" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1030.958406] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Instance network_info: |[{"id": "86406a4f-4de3-4c2d-b326-ced3cbdf893b", "address": "fa:16:3e:6e:a7:eb", "network": {"id": "1fcbe5e7-7620-481b-a739-be070f31d15f", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-84347152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e511c03cdf5c4ce9bcd9e56b425e20fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86406a4f-4d", "ovs_interfaceid": "86406a4f-4de3-4c2d-b326-ced3cbdf893b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1030.958723] env[62753]: DEBUG oslo_concurrency.lockutils [req-916b4084-08ab-4469-8ffd-cf22a492beb4 req-c407ef6c-e04f-4ba2-817a-a5d49b5a11a3 service nova] Acquired lock "refresh_cache-ff50c5f4-ab27-4f17-948e-80e2dcc623eb" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1030.958901] env[62753]: DEBUG nova.network.neutron [req-916b4084-08ab-4469-8ffd-cf22a492beb4 req-c407ef6c-e04f-4ba2-817a-a5d49b5a11a3 service nova] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Refreshing network info cache for port 86406a4f-4de3-4c2d-b326-ced3cbdf893b {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1030.960311] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:6e:a7:eb', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db00ec2e-3155-46b6-8170-082f7d86dbe7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '86406a4f-4de3-4c2d-b326-ced3cbdf893b', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1030.968025] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Creating folder: Project (e511c03cdf5c4ce9bcd9e56b425e20fb). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1030.971140] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-5a5a4410-20f1-49bf-a90b-84e841158d5a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.982753] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Created folder: Project (e511c03cdf5c4ce9bcd9e56b425e20fb) in parent group-v284541. [ 1030.982957] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Creating folder: Instances. Parent ref: group-v284666. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1030.983203] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-aa0ead41-e84b-47bd-a20f-071363bec57b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1030.992148] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Created folder: Instances in parent group-v284666. [ 1030.992372] env[62753]: DEBUG oslo.service.loopingcall [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1030.992572] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1030.993773] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-212c8d82-f044-41ce-8b20-36580c06abb0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.011321] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1031.011321] env[62753]: value = "task-1332683" [ 1031.011321] env[62753]: _type = "Task" [ 1031.011321] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.018666] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332683, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.164347] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.309s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1031.164946] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1031.167604] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.811s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1031.167838] env[62753]: DEBUG nova.objects.instance [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lazy-loading 'resources' on Instance uuid b5877d39-2f94-4a01-9db1-89a3e612095c {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1031.188550] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332680, 'name': Rename_Task, 'duration_secs': 0.145612} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.188816] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1031.189080] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d69085e6-e454-462f-af49-1f5eee9eb086 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.195313] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 1031.195313] env[62753]: value = "task-1332684" [ 1031.195313] env[62753]: _type = "Task" [ 1031.195313] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.203971] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332684, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.300108] env[62753]: DEBUG nova.network.neutron [req-916b4084-08ab-4469-8ffd-cf22a492beb4 req-c407ef6c-e04f-4ba2-817a-a5d49b5a11a3 service nova] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Updated VIF entry in instance network info cache for port 86406a4f-4de3-4c2d-b326-ced3cbdf893b. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1031.300565] env[62753]: DEBUG nova.network.neutron [req-916b4084-08ab-4469-8ffd-cf22a492beb4 req-c407ef6c-e04f-4ba2-817a-a5d49b5a11a3 service nova] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Updating instance_info_cache with network_info: [{"id": "86406a4f-4de3-4c2d-b326-ced3cbdf893b", "address": "fa:16:3e:6e:a7:eb", "network": {"id": "1fcbe5e7-7620-481b-a739-be070f31d15f", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-84347152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.9", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e511c03cdf5c4ce9bcd9e56b425e20fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap86406a4f-4d", "ovs_interfaceid": "86406a4f-4de3-4c2d-b326-ced3cbdf893b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1031.520907] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332683, 'name': CreateVM_Task, 'duration_secs': 0.300736} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.521126] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1031.521763] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1031.521933] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1031.522306] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1031.522565] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad311e39-5523-4392-93e4-1798af16aae5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.527832] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1031.527832] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dcdf36-268b-7b63-2df1-7ed40075f594" [ 1031.527832] env[62753]: _type = "Task" [ 1031.527832] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1031.535079] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dcdf36-268b-7b63-2df1-7ed40075f594, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1031.535852] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Successfully updated port: 0c86bac1-aaf9-4933-ac5a-6488554c390c {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1031.671010] env[62753]: DEBUG nova.compute.utils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1031.672441] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1031.672614] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1031.707125] env[62753]: DEBUG oslo_vmware.api [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332684, 'name': PowerOnVM_Task, 'duration_secs': 0.46732} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1031.707588] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1031.707588] env[62753]: INFO nova.compute.manager [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Took 7.75 seconds to spawn the instance on the hypervisor. [ 1031.707791] env[62753]: DEBUG nova.compute.manager [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1031.710632] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f31ff89-1c0a-4776-98f2-df508a9f67cf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.719914] env[62753]: DEBUG nova.policy [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '199d994ed55a491cb5802fa000963c85', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e511c03cdf5c4ce9bcd9e56b425e20fb', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1031.803345] env[62753]: DEBUG oslo_concurrency.lockutils [req-916b4084-08ab-4469-8ffd-cf22a492beb4 req-c407ef6c-e04f-4ba2-817a-a5d49b5a11a3 service nova] Releasing lock "refresh_cache-ff50c5f4-ab27-4f17-948e-80e2dcc623eb" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1031.873776] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-19722ab0-3751-4bd6-9131-7545952b046a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.883868] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f0719df-6fd8-458e-8c81-8ab4b91d39dd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.918325] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4800b2e-e1d4-444d-8caf-0adf02911ee7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.925900] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcac435d-a2cc-4f0a-887d-3ae3d0896f8c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1031.944993] env[62753]: DEBUG nova.compute.provider_tree [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1032.005549] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Successfully created port: 6e141f4e-b7cc-491d-b299-8ae721ee06e5 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1032.039896] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "refresh_cache-d7ba8be1-befb-4a0b-9c35-d86e7d586203" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.040088] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquired lock "refresh_cache-d7ba8be1-befb-4a0b-9c35-d86e7d586203" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.040270] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1032.041496] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52dcdf36-268b-7b63-2df1-7ed40075f594, 'name': SearchDatastore_Task, 'duration_secs': 0.009152} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.041820] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1032.042125] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1032.042413] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.042599] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1032.042838] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1032.043156] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-64e604b4-4238-4225-bdd3-04ef86fc5c69 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.053209] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1032.053453] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1032.055121] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d495e20f-e23e-4039-95d7-6ed2152e711a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.061422] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1032.061422] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52aa99cf-249c-299b-ee56-e6d5c8a32dd3" [ 1032.061422] env[62753]: _type = "Task" [ 1032.061422] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.070026] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52aa99cf-249c-299b-ee56-e6d5c8a32dd3, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.109447] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "04fd0022-f20a-4217-9e47-5381635b17a6" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.110705] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "04fd0022-f20a-4217-9e47-5381635b17a6" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.121891] env[62753]: DEBUG nova.compute.manager [req-f3869fe1-00ee-4d7d-8471-23a1ac9146b7 req-48b47381-6b9f-4922-9f81-891f6ba9da60 service nova] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Received event network-vif-plugged-0c86bac1-aaf9-4933-ac5a-6488554c390c {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1032.122272] env[62753]: DEBUG oslo_concurrency.lockutils [req-f3869fe1-00ee-4d7d-8471-23a1ac9146b7 req-48b47381-6b9f-4922-9f81-891f6ba9da60 service nova] Acquiring lock "d7ba8be1-befb-4a0b-9c35-d86e7d586203-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.122398] env[62753]: DEBUG oslo_concurrency.lockutils [req-f3869fe1-00ee-4d7d-8471-23a1ac9146b7 req-48b47381-6b9f-4922-9f81-891f6ba9da60 service nova] Lock "d7ba8be1-befb-4a0b-9c35-d86e7d586203-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.122609] env[62753]: DEBUG oslo_concurrency.lockutils [req-f3869fe1-00ee-4d7d-8471-23a1ac9146b7 req-48b47381-6b9f-4922-9f81-891f6ba9da60 service nova] Lock "d7ba8be1-befb-4a0b-9c35-d86e7d586203-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.122821] env[62753]: DEBUG nova.compute.manager [req-f3869fe1-00ee-4d7d-8471-23a1ac9146b7 req-48b47381-6b9f-4922-9f81-891f6ba9da60 service nova] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] No waiting events found dispatching network-vif-plugged-0c86bac1-aaf9-4933-ac5a-6488554c390c {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1032.123102] env[62753]: WARNING nova.compute.manager [req-f3869fe1-00ee-4d7d-8471-23a1ac9146b7 req-48b47381-6b9f-4922-9f81-891f6ba9da60 service nova] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Received unexpected event network-vif-plugged-0c86bac1-aaf9-4933-ac5a-6488554c390c for instance with vm_state building and task_state spawning. [ 1032.123369] env[62753]: DEBUG nova.compute.manager [req-f3869fe1-00ee-4d7d-8471-23a1ac9146b7 req-48b47381-6b9f-4922-9f81-891f6ba9da60 service nova] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Received event network-changed-0c86bac1-aaf9-4933-ac5a-6488554c390c {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1032.123612] env[62753]: DEBUG nova.compute.manager [req-f3869fe1-00ee-4d7d-8471-23a1ac9146b7 req-48b47381-6b9f-4922-9f81-891f6ba9da60 service nova] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Refreshing instance network info cache due to event network-changed-0c86bac1-aaf9-4933-ac5a-6488554c390c. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1032.123862] env[62753]: DEBUG oslo_concurrency.lockutils [req-f3869fe1-00ee-4d7d-8471-23a1ac9146b7 req-48b47381-6b9f-4922-9f81-891f6ba9da60 service nova] Acquiring lock "refresh_cache-d7ba8be1-befb-4a0b-9c35-d86e7d586203" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1032.180446] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1032.228776] env[62753]: INFO nova.compute.manager [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Took 18.61 seconds to build instance. [ 1032.436555] env[62753]: DEBUG oslo_concurrency.lockutils [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "c73b1ae0-dc98-47f7-babf-e96169384785" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.436844] env[62753]: DEBUG oslo_concurrency.lockutils [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "c73b1ae0-dc98-47f7-babf-e96169384785" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.437048] env[62753]: DEBUG nova.compute.manager [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Going to confirm migration 3 {{(pid=62753) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1032.449045] env[62753]: DEBUG nova.scheduler.client.report [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1032.487081] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "4d20e083-2959-453a-8875-47955bc02613" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.487456] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "4d20e083-2959-453a-8875-47955bc02613" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.487676] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "4d20e083-2959-453a-8875-47955bc02613-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1032.487870] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "4d20e083-2959-453a-8875-47955bc02613-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.488060] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "4d20e083-2959-453a-8875-47955bc02613-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.490330] env[62753]: INFO nova.compute.manager [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Terminating instance [ 1032.492061] env[62753]: DEBUG nova.compute.manager [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1032.492309] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1032.493162] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a42b371-aa66-4648-9a2d-84784a6ccad0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.500845] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1032.501084] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-4f2e69e2-20d2-4905-97eb-f558d8a79c35 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.507490] env[62753]: DEBUG oslo_vmware.api [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 1032.507490] env[62753]: value = "task-1332685" [ 1032.507490] env[62753]: _type = "Task" [ 1032.507490] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.514857] env[62753]: DEBUG oslo_vmware.api [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332685, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.571291] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52aa99cf-249c-299b-ee56-e6d5c8a32dd3, 'name': SearchDatastore_Task, 'duration_secs': 0.012297} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1032.572137] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4891193c-1797-40c9-9231-42acab9cc188 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1032.577838] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1032.577838] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5220c531-dbe0-9517-9859-132d444f32b0" [ 1032.577838] env[62753]: _type = "Task" [ 1032.577838] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1032.578699] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1032.588665] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5220c531-dbe0-9517-9859-132d444f32b0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1032.612902] env[62753]: DEBUG nova.compute.utils [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1032.722767] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Updating instance_info_cache with network_info: [{"id": "0c86bac1-aaf9-4933-ac5a-6488554c390c", "address": "fa:16:3e:38:1f:a0", "network": {"id": "1fcbe5e7-7620-481b-a739-be070f31d15f", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-84347152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e511c03cdf5c4ce9bcd9e56b425e20fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c86bac1-aa", "ovs_interfaceid": "0c86bac1-aaf9-4933-ac5a-6488554c390c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1032.730853] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9dbae57-d638-498c-aa8a-1bec1deae563 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "b1e4c8f9-471d-4cca-9a12-56191767a5b1" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.131s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.953964] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.786s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1032.956165] env[62753]: DEBUG oslo_concurrency.lockutils [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.197s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1032.956462] env[62753]: DEBUG nova.objects.instance [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lazy-loading 'resources' on Instance uuid 53a76eb0-8b0c-4e47-8387-75df4c0b191f {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1032.980859] env[62753]: INFO nova.scheduler.client.report [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Deleted allocations for instance b5877d39-2f94-4a01-9db1-89a3e612095c [ 1033.017887] env[62753]: DEBUG oslo_vmware.api [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332685, 'name': PowerOffVM_Task, 'duration_secs': 0.186391} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.018997] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1033.019146] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1033.020798] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b32eec88-236e-4043-bd4f-806c3a900ff2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.027950] env[62753]: DEBUG oslo_concurrency.lockutils [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.028167] env[62753]: DEBUG oslo_concurrency.lockutils [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.028349] env[62753]: DEBUG nova.network.neutron [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1033.028536] env[62753]: DEBUG nova.objects.instance [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lazy-loading 'info_cache' on Instance uuid c73b1ae0-dc98-47f7-babf-e96169384785 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.091042] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5220c531-dbe0-9517-9859-132d444f32b0, 'name': SearchDatastore_Task, 'duration_secs': 0.010463} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1033.091320] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.091592] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] ff50c5f4-ab27-4f17-948e-80e2dcc623eb/ff50c5f4-ab27-4f17-948e-80e2dcc623eb.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1033.091870] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-71dcdeb2-2357-4c67-95e5-9ef076cc2d3e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.098565] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1033.098565] env[62753]: value = "task-1332687" [ 1033.098565] env[62753]: _type = "Task" [ 1033.098565] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.106744] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332687, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.115601] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "04fd0022-f20a-4217-9e47-5381635b17a6" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.192019] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1033.218800] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1033.218800] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1033.218996] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1033.219906] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1033.219906] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1033.219906] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1033.219906] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1033.220156] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1033.220326] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1033.220578] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1033.220809] env[62753]: DEBUG nova.virt.hardware [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1033.221801] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11061b7d-0c3c-4c6c-b7b7-007d6f36fdfd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.225084] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Releasing lock "refresh_cache-d7ba8be1-befb-4a0b-9c35-d86e7d586203" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1033.225446] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Instance network_info: |[{"id": "0c86bac1-aaf9-4933-ac5a-6488554c390c", "address": "fa:16:3e:38:1f:a0", "network": {"id": "1fcbe5e7-7620-481b-a739-be070f31d15f", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-84347152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e511c03cdf5c4ce9bcd9e56b425e20fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c86bac1-aa", "ovs_interfaceid": "0c86bac1-aaf9-4933-ac5a-6488554c390c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1033.225753] env[62753]: DEBUG oslo_concurrency.lockutils [req-f3869fe1-00ee-4d7d-8471-23a1ac9146b7 req-48b47381-6b9f-4922-9f81-891f6ba9da60 service nova] Acquired lock "refresh_cache-d7ba8be1-befb-4a0b-9c35-d86e7d586203" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.225960] env[62753]: DEBUG nova.network.neutron [req-f3869fe1-00ee-4d7d-8471-23a1ac9146b7 req-48b47381-6b9f-4922-9f81-891f6ba9da60 service nova] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Refreshing network info cache for port 0c86bac1-aaf9-4933-ac5a-6488554c390c {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1033.227283] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:38:1f:a0', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db00ec2e-3155-46b6-8170-082f7d86dbe7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0c86bac1-aaf9-4933-ac5a-6488554c390c', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1033.237308] env[62753]: DEBUG oslo.service.loopingcall [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1033.241976] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1033.245226] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3d3c4cba-9f56-4919-beb0-bef7ac93770b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.262431] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8208de1-865b-4350-b0e5-2d1776ee33a5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.266383] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1033.266608] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1033.266812] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Deleting the datastore file [datastore2] 4d20e083-2959-453a-8875-47955bc02613 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1033.267708] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4ea81212-30cc-48ad-ab14-803d7e53b10b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.271609] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1033.271609] env[62753]: value = "task-1332688" [ 1033.271609] env[62753]: _type = "Task" [ 1033.271609] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.284019] env[62753]: DEBUG oslo_vmware.api [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for the task: (returnval){ [ 1033.284019] env[62753]: value = "task-1332689" [ 1033.284019] env[62753]: _type = "Task" [ 1033.284019] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1033.290743] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332688, 'name': CreateVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.295311] env[62753]: DEBUG oslo_vmware.api [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332689, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.489423] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b029556d-8575-4559-8355-173c391728ef tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "b5877d39-2f94-4a01-9db1-89a3e612095c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.028s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1033.523808] env[62753]: DEBUG oslo_concurrency.lockutils [None req-edebb745-36f7-44d1-be0f-beafcaa2c326 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "b1e4c8f9-471d-4cca-9a12-56191767a5b1" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1033.524144] env[62753]: DEBUG oslo_concurrency.lockutils [None req-edebb745-36f7-44d1-be0f-beafcaa2c326 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "b1e4c8f9-471d-4cca-9a12-56191767a5b1" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1033.524297] env[62753]: DEBUG nova.compute.manager [None req-edebb745-36f7-44d1-be0f-beafcaa2c326 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1033.525144] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9b88974-20f9-465b-a424-83a671722096 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.533785] env[62753]: DEBUG nova.compute.manager [None req-edebb745-36f7-44d1-be0f-beafcaa2c326 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62753) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1033.534411] env[62753]: DEBUG nova.objects.instance [None req-edebb745-36f7-44d1-be0f-beafcaa2c326 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lazy-loading 'flavor' on Instance uuid b1e4c8f9-471d-4cca-9a12-56191767a5b1 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1033.609076] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332687, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.634306] env[62753]: DEBUG nova.network.neutron [req-f3869fe1-00ee-4d7d-8471-23a1ac9146b7 req-48b47381-6b9f-4922-9f81-891f6ba9da60 service nova] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Updated VIF entry in instance network info cache for port 0c86bac1-aaf9-4933-ac5a-6488554c390c. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1033.635119] env[62753]: DEBUG nova.network.neutron [req-f3869fe1-00ee-4d7d-8471-23a1ac9146b7 req-48b47381-6b9f-4922-9f81-891f6ba9da60 service nova] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Updating instance_info_cache with network_info: [{"id": "0c86bac1-aaf9-4933-ac5a-6488554c390c", "address": "fa:16:3e:38:1f:a0", "network": {"id": "1fcbe5e7-7620-481b-a739-be070f31d15f", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-84347152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e511c03cdf5c4ce9bcd9e56b425e20fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0c86bac1-aa", "ovs_interfaceid": "0c86bac1-aaf9-4933-ac5a-6488554c390c", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1033.688402] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea99ae40-ab60-4e90-bd0f-4eac01e831a2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.698475] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0785e83e-09bb-461d-84e3-942cf24682a9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.709694] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Successfully updated port: 6e141f4e-b7cc-491d-b299-8ae721ee06e5 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1033.738587] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "refresh_cache-961ced74-10af-4443-857e-545cd890674e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1033.738906] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquired lock "refresh_cache-961ced74-10af-4443-857e-545cd890674e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1033.738906] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1033.740359] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bfd6a838-f1a0-4860-98e1-92135af8ac67 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.749376] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e256e419-8909-4d69-883a-96ace71cc7a9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1033.763208] env[62753]: DEBUG nova.compute.provider_tree [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1033.792551] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332688, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1033.795355] env[62753]: DEBUG oslo_vmware.api [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332689, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.044535] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-edebb745-36f7-44d1-be0f-beafcaa2c326 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1034.044535] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d615be0a-d07a-4932-ad28-7fc4b48df7a8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.050901] env[62753]: DEBUG oslo_vmware.api [None req-edebb745-36f7-44d1-be0f-beafcaa2c326 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 1034.050901] env[62753]: value = "task-1332690" [ 1034.050901] env[62753]: _type = "Task" [ 1034.050901] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.059380] env[62753]: DEBUG oslo_vmware.api [None req-edebb745-36f7-44d1-be0f-beafcaa2c326 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332690, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.110580] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332687, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.57337} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.110905] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] ff50c5f4-ab27-4f17-948e-80e2dcc623eb/ff50c5f4-ab27-4f17-948e-80e2dcc623eb.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1034.111152] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1034.111412] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-66732170-bdf3-485f-9993-8ae5e3e60615 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.117784] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1034.117784] env[62753]: value = "task-1332691" [ 1034.117784] env[62753]: _type = "Task" [ 1034.117784] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.131288] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332691, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.137073] env[62753]: DEBUG oslo_concurrency.lockutils [req-f3869fe1-00ee-4d7d-8471-23a1ac9146b7 req-48b47381-6b9f-4922-9f81-891f6ba9da60 service nova] Releasing lock "refresh_cache-d7ba8be1-befb-4a0b-9c35-d86e7d586203" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.212084] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "04fd0022-f20a-4217-9e47-5381635b17a6" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.213077] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "04fd0022-f20a-4217-9e47-5381635b17a6" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.213077] env[62753]: INFO nova.compute.manager [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Attaching volume e93861cc-b07a-4254-b33d-0814321eef34 to /dev/sdb [ 1034.252823] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8339c2be-eb31-4776-af63-7d84e0dac6a6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.267659] env[62753]: DEBUG nova.scheduler.client.report [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1034.275226] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-46be4b9f-913b-486a-9360-59c8406c1ff0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.280565] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1034.289332] env[62753]: DEBUG nova.virt.block_device [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Updating existing volume attachment record: f4aee6f7-31a7-48aa-9aa2-ab14f8ee51b9 {{(pid=62753) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1034.306664] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332688, 'name': CreateVM_Task, 'duration_secs': 0.867108} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.314961] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1034.316097] env[62753]: DEBUG oslo_vmware.api [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Task: {'id': task-1332689, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.70551} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.316740] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.316900] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.317240] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1034.317881] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1034.317964] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1034.318146] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1034.318328] env[62753]: INFO nova.compute.manager [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] [instance: 4d20e083-2959-453a-8875-47955bc02613] Took 1.83 seconds to destroy the instance on the hypervisor. [ 1034.318552] env[62753]: DEBUG oslo.service.loopingcall [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1034.318747] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e2346476-6443-4aa7-8222-5359a657df5d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.320441] env[62753]: DEBUG nova.compute.manager [-] [instance: 4d20e083-2959-453a-8875-47955bc02613] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1034.320547] env[62753]: DEBUG nova.network.neutron [-] [instance: 4d20e083-2959-453a-8875-47955bc02613] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1034.325017] env[62753]: DEBUG nova.compute.manager [req-ac4a9f92-376d-4afa-887a-761ee75f9542 req-48bebfc8-7f3e-4b7c-8701-530f0f96c874 service nova] [instance: 961ced74-10af-4443-857e-545cd890674e] Received event network-vif-plugged-6e141f4e-b7cc-491d-b299-8ae721ee06e5 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1034.325225] env[62753]: DEBUG oslo_concurrency.lockutils [req-ac4a9f92-376d-4afa-887a-761ee75f9542 req-48bebfc8-7f3e-4b7c-8701-530f0f96c874 service nova] Acquiring lock "961ced74-10af-4443-857e-545cd890674e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1034.325454] env[62753]: DEBUG oslo_concurrency.lockutils [req-ac4a9f92-376d-4afa-887a-761ee75f9542 req-48bebfc8-7f3e-4b7c-8701-530f0f96c874 service nova] Lock "961ced74-10af-4443-857e-545cd890674e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.325630] env[62753]: DEBUG oslo_concurrency.lockutils [req-ac4a9f92-376d-4afa-887a-761ee75f9542 req-48bebfc8-7f3e-4b7c-8701-530f0f96c874 service nova] Lock "961ced74-10af-4443-857e-545cd890674e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.325900] env[62753]: DEBUG nova.compute.manager [req-ac4a9f92-376d-4afa-887a-761ee75f9542 req-48bebfc8-7f3e-4b7c-8701-530f0f96c874 service nova] [instance: 961ced74-10af-4443-857e-545cd890674e] No waiting events found dispatching network-vif-plugged-6e141f4e-b7cc-491d-b299-8ae721ee06e5 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1034.325983] env[62753]: WARNING nova.compute.manager [req-ac4a9f92-376d-4afa-887a-761ee75f9542 req-48bebfc8-7f3e-4b7c-8701-530f0f96c874 service nova] [instance: 961ced74-10af-4443-857e-545cd890674e] Received unexpected event network-vif-plugged-6e141f4e-b7cc-491d-b299-8ae721ee06e5 for instance with vm_state building and task_state spawning. [ 1034.326136] env[62753]: DEBUG nova.compute.manager [req-ac4a9f92-376d-4afa-887a-761ee75f9542 req-48bebfc8-7f3e-4b7c-8701-530f0f96c874 service nova] [instance: 961ced74-10af-4443-857e-545cd890674e] Received event network-changed-6e141f4e-b7cc-491d-b299-8ae721ee06e5 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1034.326284] env[62753]: DEBUG nova.compute.manager [req-ac4a9f92-376d-4afa-887a-761ee75f9542 req-48bebfc8-7f3e-4b7c-8701-530f0f96c874 service nova] [instance: 961ced74-10af-4443-857e-545cd890674e] Refreshing instance network info cache due to event network-changed-6e141f4e-b7cc-491d-b299-8ae721ee06e5. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1034.326453] env[62753]: DEBUG oslo_concurrency.lockutils [req-ac4a9f92-376d-4afa-887a-761ee75f9542 req-48bebfc8-7f3e-4b7c-8701-530f0f96c874 service nova] Acquiring lock "refresh_cache-961ced74-10af-4443-857e-545cd890674e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.328358] env[62753]: DEBUG nova.network.neutron [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating instance_info_cache with network_info: [{"id": "81591716-fb4a-427f-b3e9-a8b265fe1ce1", "address": "fa:16:3e:3e:ef:2c", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81591716-fb", "ovs_interfaceid": "81591716-fb4a-427f-b3e9-a8b265fe1ce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.333224] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1034.333224] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e529b8-153f-8ea1-3277-1b9b98dec20d" [ 1034.333224] env[62753]: _type = "Task" [ 1034.333224] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.341699] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e529b8-153f-8ea1-3277-1b9b98dec20d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.537632] env[62753]: DEBUG nova.network.neutron [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Updating instance_info_cache with network_info: [{"id": "6e141f4e-b7cc-491d-b299-8ae721ee06e5", "address": "fa:16:3e:29:b2:58", "network": {"id": "1fcbe5e7-7620-481b-a739-be070f31d15f", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-84347152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e511c03cdf5c4ce9bcd9e56b425e20fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e141f4e-b7", "ovs_interfaceid": "6e141f4e-b7cc-491d-b299-8ae721ee06e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1034.560390] env[62753]: DEBUG oslo_vmware.api [None req-edebb745-36f7-44d1-be0f-beafcaa2c326 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332690, 'name': PowerOffVM_Task, 'duration_secs': 0.213189} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.560658] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-edebb745-36f7-44d1-be0f-beafcaa2c326 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1034.560837] env[62753]: DEBUG nova.compute.manager [None req-edebb745-36f7-44d1-be0f-beafcaa2c326 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1034.561612] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f07c581b-fb1f-428e-98e8-47cc0b95b812 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.626653] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332691, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.075691} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.627836] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1034.627932] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cbae0d02-23c4-415c-bcaf-d0ad3b11815b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.650497] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Reconfiguring VM instance instance-0000005e to attach disk [datastore2] ff50c5f4-ab27-4f17-948e-80e2dcc623eb/ff50c5f4-ab27-4f17-948e-80e2dcc623eb.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1034.650844] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ba26e394-a465-4729-80ae-1eba0a96cc30 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.672051] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1034.672051] env[62753]: value = "task-1332695" [ 1034.672051] env[62753]: _type = "Task" [ 1034.672051] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.680854] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332695, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1034.773010] env[62753]: DEBUG oslo_concurrency.lockutils [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.817s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1034.776024] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.852s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1034.776979] env[62753]: INFO nova.compute.claims [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1034.793287] env[62753]: INFO nova.scheduler.client.report [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Deleted allocations for instance 53a76eb0-8b0c-4e47-8387-75df4c0b191f [ 1034.831644] env[62753]: DEBUG oslo_concurrency.lockutils [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.831909] env[62753]: DEBUG nova.objects.instance [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lazy-loading 'migration_context' on Instance uuid c73b1ae0-dc98-47f7-babf-e96169384785 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1034.843402] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e529b8-153f-8ea1-3277-1b9b98dec20d, 'name': SearchDatastore_Task, 'duration_secs': 0.008635} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1034.843683] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1034.843908] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1034.844155] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1034.844368] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1034.844560] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1034.845033] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-469e661a-4450-4f2e-8aab-522eea355a53 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.853457] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1034.853688] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1034.854509] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81c5e2ef-3d6c-404d-be45-a18cae22338f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1034.861118] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1034.861118] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5253ff6c-bd3f-0846-eb56-bc1f5231dbd9" [ 1034.861118] env[62753]: _type = "Task" [ 1034.861118] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1034.869733] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5253ff6c-bd3f-0846-eb56-bc1f5231dbd9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.040251] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Releasing lock "refresh_cache-961ced74-10af-4443-857e-545cd890674e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.040594] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Instance network_info: |[{"id": "6e141f4e-b7cc-491d-b299-8ae721ee06e5", "address": "fa:16:3e:29:b2:58", "network": {"id": "1fcbe5e7-7620-481b-a739-be070f31d15f", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-84347152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e511c03cdf5c4ce9bcd9e56b425e20fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e141f4e-b7", "ovs_interfaceid": "6e141f4e-b7cc-491d-b299-8ae721ee06e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1035.040917] env[62753]: DEBUG oslo_concurrency.lockutils [req-ac4a9f92-376d-4afa-887a-761ee75f9542 req-48bebfc8-7f3e-4b7c-8701-530f0f96c874 service nova] Acquired lock "refresh_cache-961ced74-10af-4443-857e-545cd890674e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.041120] env[62753]: DEBUG nova.network.neutron [req-ac4a9f92-376d-4afa-887a-761ee75f9542 req-48bebfc8-7f3e-4b7c-8701-530f0f96c874 service nova] [instance: 961ced74-10af-4443-857e-545cd890674e] Refreshing network info cache for port 6e141f4e-b7cc-491d-b299-8ae721ee06e5 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1035.042349] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:29:b2:58', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'db00ec2e-3155-46b6-8170-082f7d86dbe7', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '6e141f4e-b7cc-491d-b299-8ae721ee06e5', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1035.049869] env[62753]: DEBUG oslo.service.loopingcall [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1035.052779] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 961ced74-10af-4443-857e-545cd890674e] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1035.053275] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-9cb3b870-4e4a-43df-a283-2e2e313f7ab4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.073366] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.073586] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1035.075020] env[62753]: DEBUG oslo_concurrency.lockutils [None req-edebb745-36f7-44d1-be0f-beafcaa2c326 tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "b1e4c8f9-471d-4cca-9a12-56191767a5b1" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.551s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.077368] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1035.077368] env[62753]: value = "task-1332696" [ 1035.077368] env[62753]: _type = "Task" [ 1035.077368] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.086960] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332696, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.167349] env[62753]: DEBUG nova.network.neutron [-] [instance: 4d20e083-2959-453a-8875-47955bc02613] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.182393] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332695, 'name': ReconfigVM_Task, 'duration_secs': 0.284189} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.182691] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Reconfigured VM instance instance-0000005e to attach disk [datastore2] ff50c5f4-ab27-4f17-948e-80e2dcc623eb/ff50c5f4-ab27-4f17-948e-80e2dcc623eb.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1035.185289] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-24c74bc9-a099-49dd-8bf2-e59ec7556af2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.192615] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1035.192615] env[62753]: value = "task-1332697" [ 1035.192615] env[62753]: _type = "Task" [ 1035.192615] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.200914] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332697, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.300506] env[62753]: DEBUG oslo_concurrency.lockutils [None req-387004a4-5cea-4342-8cc3-a2a24128d100 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "53a76eb0-8b0c-4e47-8387-75df4c0b191f" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.630s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1035.302014] env[62753]: DEBUG nova.network.neutron [req-ac4a9f92-376d-4afa-887a-761ee75f9542 req-48bebfc8-7f3e-4b7c-8701-530f0f96c874 service nova] [instance: 961ced74-10af-4443-857e-545cd890674e] Updated VIF entry in instance network info cache for port 6e141f4e-b7cc-491d-b299-8ae721ee06e5. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1035.302379] env[62753]: DEBUG nova.network.neutron [req-ac4a9f92-376d-4afa-887a-761ee75f9542 req-48bebfc8-7f3e-4b7c-8701-530f0f96c874 service nova] [instance: 961ced74-10af-4443-857e-545cd890674e] Updating instance_info_cache with network_info: [{"id": "6e141f4e-b7cc-491d-b299-8ae721ee06e5", "address": "fa:16:3e:29:b2:58", "network": {"id": "1fcbe5e7-7620-481b-a739-be070f31d15f", "bridge": "br-int", "label": "tempest-ListServersNegativeTestJSON-84347152-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e511c03cdf5c4ce9bcd9e56b425e20fb", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "db00ec2e-3155-46b6-8170-082f7d86dbe7", "external-id": "nsx-vlan-transportzone-332", "segmentation_id": 332, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap6e141f4e-b7", "ovs_interfaceid": "6e141f4e-b7cc-491d-b299-8ae721ee06e5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1035.334481] env[62753]: DEBUG nova.objects.base [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1035.335483] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11d89e2f-d11b-4529-adb4-76d99269e08a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.354936] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2a899b6-7879-47a3-800e-3aefbe711ca8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.360368] env[62753]: DEBUG oslo_vmware.api [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1035.360368] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523fa660-8541-a998-e097-21fd94edb2a8" [ 1035.360368] env[62753]: _type = "Task" [ 1035.360368] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.371105] env[62753]: DEBUG oslo_vmware.api [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523fa660-8541-a998-e097-21fd94edb2a8, 'name': SearchDatastore_Task, 'duration_secs': 0.00826} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.374454] env[62753]: DEBUG oslo_concurrency.lockutils [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1035.374725] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5253ff6c-bd3f-0846-eb56-bc1f5231dbd9, 'name': SearchDatastore_Task, 'duration_secs': 0.009402} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.375443] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5f638072-ecf7-486d-b22c-db9ea9acd1d3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.380189] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1035.380189] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5262ea8a-769a-548c-2249-d0a9bef51478" [ 1035.380189] env[62753]: _type = "Task" [ 1035.380189] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.387604] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5262ea8a-769a-548c-2249-d0a9bef51478, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.577952] env[62753]: DEBUG nova.compute.manager [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1035.594015] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332696, 'name': CreateVM_Task, 'duration_secs': 0.445794} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.594015] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 961ced74-10af-4443-857e-545cd890674e] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1035.594776] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1035.595045] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1035.595474] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1035.596108] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5d9476b4-d43c-466e-93ca-85bf4f9add26 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.600743] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1035.600743] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f5a265-6268-f3f8-15ff-aaee364bf1a9" [ 1035.600743] env[62753]: _type = "Task" [ 1035.600743] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.609090] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f5a265-6268-f3f8-15ff-aaee364bf1a9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.669808] env[62753]: INFO nova.compute.manager [-] [instance: 4d20e083-2959-453a-8875-47955bc02613] Took 1.35 seconds to deallocate network for instance. [ 1035.702483] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332697, 'name': Rename_Task, 'duration_secs': 0.136351} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.703382] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1035.703636] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6beda345-c447-4c69-b347-4b2925338874 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.710530] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1035.710530] env[62753]: value = "task-1332698" [ 1035.710530] env[62753]: _type = "Task" [ 1035.710530] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.718069] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332698, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.804874] env[62753]: DEBUG oslo_concurrency.lockutils [req-ac4a9f92-376d-4afa-887a-761ee75f9542 req-48bebfc8-7f3e-4b7c-8701-530f0f96c874 service nova] Releasing lock "refresh_cache-961ced74-10af-4443-857e-545cd890674e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.893451] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5262ea8a-769a-548c-2249-d0a9bef51478, 'name': SearchDatastore_Task, 'duration_secs': 0.008365} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1035.893723] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1035.893987] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] d7ba8be1-befb-4a0b-9c35-d86e7d586203/d7ba8be1-befb-4a0b-9c35-d86e7d586203.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1035.894314] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2067314e-dd1e-4de3-ad20-3d4b2face8b6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.904410] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1035.904410] env[62753]: value = "task-1332699" [ 1035.904410] env[62753]: _type = "Task" [ 1035.904410] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1035.914387] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332699, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1035.981955] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a20aa557-ff64-4d69-8ab4-7e7277ad9b45 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1035.989737] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87f7059b-931f-422a-95f5-7a42f8332173 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.020518] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82edae9c-f4f1-44e8-ac9f-97386d6c0fd4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.028149] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cc3af6d8-fbba-4a94-8fd9-8f87a3f9ae46 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.042531] env[62753]: DEBUG nova.compute.provider_tree [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1036.110797] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.119561] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f5a265-6268-f3f8-15ff-aaee364bf1a9, 'name': SearchDatastore_Task, 'duration_secs': 0.009181} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.120053] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1036.120463] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1036.120838] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1036.121099] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1036.121405] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1036.121845] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3d262ee2-de1d-4662-bfac-bb2427ff668a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.140574] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1036.140827] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1036.141634] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-238854b9-28dc-41ad-90aa-8de10a834437 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.147929] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1036.147929] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f85ae2-3a41-2ece-9c80-38fefd61412b" [ 1036.147929] env[62753]: _type = "Task" [ 1036.147929] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.157100] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f85ae2-3a41-2ece-9c80-38fefd61412b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.176250] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.221219] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332698, 'name': PowerOnVM_Task, 'duration_secs': 0.468065} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.221498] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1036.221704] env[62753]: INFO nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Took 7.65 seconds to spawn the instance on the hypervisor. [ 1036.221890] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1036.222937] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7a2795a-a5c8-4b39-8d08-eb1696d503e3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.234869] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "b1e4c8f9-471d-4cca-9a12-56191767a5b1" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.235229] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "b1e4c8f9-471d-4cca-9a12-56191767a5b1" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.235550] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "b1e4c8f9-471d-4cca-9a12-56191767a5b1-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.235804] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "b1e4c8f9-471d-4cca-9a12-56191767a5b1-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.236083] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "b1e4c8f9-471d-4cca-9a12-56191767a5b1-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.238435] env[62753]: INFO nova.compute.manager [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Terminating instance [ 1036.240310] env[62753]: DEBUG nova.compute.manager [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1036.240508] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1036.241384] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fbd7b8f8-8b2a-485f-8333-b157aa128696 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.249471] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1036.249800] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a7dcb27f-e043-4be3-aefb-38ac634db8b1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.338819] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1036.339131] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1036.339434] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Deleting the datastore file [datastore2] b1e4c8f9-471d-4cca-9a12-56191767a5b1 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1036.339585] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-301c06ec-b276-4a6f-888d-3a7b99e75ff4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.346802] env[62753]: DEBUG oslo_vmware.api [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for the task: (returnval){ [ 1036.346802] env[62753]: value = "task-1332701" [ 1036.346802] env[62753]: _type = "Task" [ 1036.346802] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.352672] env[62753]: DEBUG nova.compute.manager [req-245125f9-8f0a-4b9d-8aa8-b9696fa173c5 req-bfea5c8e-9f73-42fe-adb9-889bf4acb898 service nova] [instance: 4d20e083-2959-453a-8875-47955bc02613] Received event network-vif-deleted-9c52b1e4-82d2-4449-91f3-a624596bdc79 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1036.359700] env[62753]: DEBUG oslo_vmware.api [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332701, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.406573] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "c3679c66-2763-4a04-8d14-e103bd952798" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.406573] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "c3679c66-2763-4a04-8d14-e103bd952798" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.406573] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "c3679c66-2763-4a04-8d14-e103bd952798-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1036.406573] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "c3679c66-2763-4a04-8d14-e103bd952798-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1036.406912] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "c3679c66-2763-4a04-8d14-e103bd952798-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1036.412797] env[62753]: INFO nova.compute.manager [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Terminating instance [ 1036.415403] env[62753]: DEBUG nova.compute.manager [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1036.415682] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1036.416594] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d492af92-c444-419a-accf-cd58a90c0625 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.423233] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332699, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.427717] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1036.428025] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54776475-0da7-4bca-bc15-12360b41269b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.434149] env[62753]: DEBUG oslo_vmware.api [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 1036.434149] env[62753]: value = "task-1332702" [ 1036.434149] env[62753]: _type = "Task" [ 1036.434149] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.442338] env[62753]: DEBUG oslo_vmware.api [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332702, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.545918] env[62753]: DEBUG nova.scheduler.client.report [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1036.658416] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f85ae2-3a41-2ece-9c80-38fefd61412b, 'name': SearchDatastore_Task, 'duration_secs': 0.065414} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.659196] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2dc6c234-b52f-4ac7-b837-56867e96284c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1036.664714] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1036.664714] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c7b01a-d9fa-8415-ed7d-5c8c9ce4b942" [ 1036.664714] env[62753]: _type = "Task" [ 1036.664714] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1036.673964] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c7b01a-d9fa-8415-ed7d-5c8c9ce4b942, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1036.743960] env[62753]: INFO nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Took 16.04 seconds to build instance. [ 1036.857696] env[62753]: DEBUG oslo_vmware.api [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Task: {'id': task-1332701, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.417286} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1036.858101] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1036.858214] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1036.858446] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1036.858565] env[62753]: INFO nova.compute.manager [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Took 0.62 seconds to destroy the instance on the hypervisor. [ 1036.858815] env[62753]: DEBUG oslo.service.loopingcall [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1036.859020] env[62753]: DEBUG nova.compute.manager [-] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1036.859119] env[62753]: DEBUG nova.network.neutron [-] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1036.920288] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332699, 'name': CopyVirtualDisk_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.603220] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.827s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.603540] env[62753]: DEBUG nova.compute.manager [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1037.610018] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "ff50c5f4-ab27-4f17-948e-80e2dcc623eb" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.911s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1037.610018] env[62753]: DEBUG oslo_vmware.api [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332702, 'name': PowerOffVM_Task, 'duration_secs': 0.188943} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.610018] env[62753]: DEBUG oslo_concurrency.lockutils [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 8.837s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1037.610018] env[62753]: DEBUG nova.objects.instance [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lazy-loading 'resources' on Instance uuid 9d7066e1-3721-42d2-8e80-91eacbbd8b80 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1037.614976] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1037.615165] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1037.615686] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3d2c0803-35b8-41cc-949d-a5e28b150585 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.623834] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c7b01a-d9fa-8415-ed7d-5c8c9ce4b942, 'name': SearchDatastore_Task, 'duration_secs': 0.045633} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.626848] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1037.627108] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 961ced74-10af-4443-857e-545cd890674e/961ced74-10af-4443-857e-545cd890674e.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1037.627594] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332699, 'name': CopyVirtualDisk_Task, 'duration_secs': 1.551553} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1037.627779] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-6bc3cfac-089b-43eb-8d7a-b14145f27cec {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.629733] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] d7ba8be1-befb-4a0b-9c35-d86e7d586203/d7ba8be1-befb-4a0b-9c35-d86e7d586203.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1037.629733] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1037.630140] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-21afe463-0a1a-4860-a2f4-556a1ff778c9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.639023] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1037.639023] env[62753]: value = "task-1332706" [ 1037.639023] env[62753]: _type = "Task" [ 1037.639023] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.639023] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1037.639023] env[62753]: value = "task-1332705" [ 1037.639023] env[62753]: _type = "Task" [ 1037.639023] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.650446] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332706, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.653263] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332705, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1037.690849] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1037.691131] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1037.691347] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Deleting the datastore file [datastore2] c3679c66-2763-4a04-8d14-e103bd952798 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1037.691676] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-e541c447-cf49-4611-822c-bb99a93cbfd2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1037.700380] env[62753]: DEBUG oslo_vmware.api [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for the task: (returnval){ [ 1037.700380] env[62753]: value = "task-1332707" [ 1037.700380] env[62753]: _type = "Task" [ 1037.700380] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1037.710152] env[62753]: DEBUG oslo_vmware.api [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332707, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.109884] env[62753]: DEBUG nova.compute.utils [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1038.111739] env[62753]: DEBUG nova.compute.manager [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1038.112118] env[62753]: DEBUG nova.network.neutron [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1038.119212] env[62753]: DEBUG nova.network.neutron [-] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1038.154071] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332706, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.076949} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.157341] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1038.158063] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332705, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.160877] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e2fa0ec-1921-474f-8f57-ad9b2050e2a7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.164697] env[62753]: DEBUG nova.policy [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '9b6207815e8041b4bba3069635d7d962', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': 'e6c8a6bc946a469fa85cc3ab80d7333a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1038.187726] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Reconfiguring VM instance instance-0000005f to attach disk [datastore2] d7ba8be1-befb-4a0b-9c35-d86e7d586203/d7ba8be1-befb-4a0b-9c35-d86e7d586203.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1038.190703] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-160e6231-4310-45c8-8afe-93d48c7c0630 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.214179] env[62753]: DEBUG oslo_vmware.api [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Task: {'id': task-1332707, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.123419} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.215519] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1038.215728] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1038.215907] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1038.216096] env[62753]: INFO nova.compute.manager [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Took 1.80 seconds to destroy the instance on the hypervisor. [ 1038.216360] env[62753]: DEBUG oslo.service.loopingcall [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1038.216648] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1038.216648] env[62753]: value = "task-1332708" [ 1038.216648] env[62753]: _type = "Task" [ 1038.216648] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.219164] env[62753]: DEBUG nova.compute.manager [-] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1038.219269] env[62753]: DEBUG nova.network.neutron [-] [instance: c3679c66-2763-4a04-8d14-e103bd952798] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1038.231147] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332708, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.366291] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0f9762f1-e95d-47db-94fe-bf9623073557 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.374633] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be9764c8-c0f1-4e9e-90f0-abe57beab5b3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.380707] env[62753]: DEBUG nova.compute.manager [req-6ee6ee64-eef3-4fdf-a1d4-f6b617d0f58a req-04ee3621-d274-4c85-a8bc-65cf677546b0 service nova] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Received event network-vif-deleted-b3e995ee-b33d-47b0-8302-8171441ddfd6 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1038.407227] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6727de73-f69b-4008-8e42-dbdee49b9175 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.415029] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0c0d52d-31ed-4e25-8bfb-73b1bfe8c2e7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.429099] env[62753]: DEBUG nova.compute.provider_tree [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1038.579689] env[62753]: DEBUG nova.network.neutron [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Successfully created port: 31af6b98-8cec-4501-9fd5-26eda99426d4 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1038.616165] env[62753]: DEBUG nova.compute.manager [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1038.626034] env[62753]: INFO nova.compute.manager [-] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Took 1.77 seconds to deallocate network for instance. [ 1038.655924] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332705, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.546129} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.656211] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 961ced74-10af-4443-857e-545cd890674e/961ced74-10af-4443-857e-545cd890674e.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1038.656430] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1038.656689] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b1fa0575-bd5f-4539-bf01-332fe54ac136 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.663379] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1038.663379] env[62753]: value = "task-1332709" [ 1038.663379] env[62753]: _type = "Task" [ 1038.663379] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.672374] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332709, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.732720] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332708, 'name': ReconfigVM_Task, 'duration_secs': 0.468421} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1038.733391] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Reconfigured VM instance instance-0000005f to attach disk [datastore2] d7ba8be1-befb-4a0b-9c35-d86e7d586203/d7ba8be1-befb-4a0b-9c35-d86e7d586203.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1038.733759] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-8b0e8c45-7ff7-4c0d-bc55-780cd3a2b312 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.739447] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1038.739447] env[62753]: value = "task-1332710" [ 1038.739447] env[62753]: _type = "Task" [ 1038.739447] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.749383] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332710, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.848110] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Volume attach. Driver type: vmdk {{(pid=62753) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1038.848438] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284671', 'volume_id': 'e93861cc-b07a-4254-b33d-0814321eef34', 'name': 'volume-e93861cc-b07a-4254-b33d-0814321eef34', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '04fd0022-f20a-4217-9e47-5381635b17a6', 'attached_at': '', 'detached_at': '', 'volume_id': 'e93861cc-b07a-4254-b33d-0814321eef34', 'serial': 'e93861cc-b07a-4254-b33d-0814321eef34'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1038.849400] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f371641-e7c8-45bc-b8f4-7477452fd52b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.867708] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-57667a45-b6d5-4750-ae81-5076d203905f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.898029] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Reconfiguring VM instance instance-00000058 to attach disk [datastore1] volume-e93861cc-b07a-4254-b33d-0814321eef34/volume-e93861cc-b07a-4254-b33d-0814321eef34.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1038.898410] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9fb1e69f-c90e-47ad-afbb-89f7d2932f35 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1038.916913] env[62753]: DEBUG oslo_vmware.api [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1038.916913] env[62753]: value = "task-1332711" [ 1038.916913] env[62753]: _type = "Task" [ 1038.916913] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1038.924625] env[62753]: DEBUG oslo_vmware.api [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332711, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1038.932245] env[62753]: DEBUG nova.scheduler.client.report [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1039.124778] env[62753]: DEBUG nova.network.neutron [-] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1039.136101] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1039.173591] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332709, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.068498} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.173591] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1039.174328] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c4c82a7-981c-43f7-ae43-e3c17c403003 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.198596] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Reconfiguring VM instance instance-00000060 to attach disk [datastore2] 961ced74-10af-4443-857e-545cd890674e/961ced74-10af-4443-857e-545cd890674e.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1039.199024] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-97749b4c-1a88-40f2-91d2-211e4d9380a2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.218799] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1039.218799] env[62753]: value = "task-1332712" [ 1039.218799] env[62753]: _type = "Task" [ 1039.218799] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.227572] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332712, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.247983] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332710, 'name': Rename_Task, 'duration_secs': 0.174619} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.248293] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1039.248533] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1346cd2c-9905-4874-b2d0-0beb44a28051 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.254275] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1039.254275] env[62753]: value = "task-1332713" [ 1039.254275] env[62753]: _type = "Task" [ 1039.254275] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.263163] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332713, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.426902] env[62753]: DEBUG oslo_vmware.api [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332711, 'name': ReconfigVM_Task, 'duration_secs': 0.407303} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.427214] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Reconfigured VM instance instance-00000058 to attach disk [datastore1] volume-e93861cc-b07a-4254-b33d-0814321eef34/volume-e93861cc-b07a-4254-b33d-0814321eef34.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1039.431730] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5cbe34fa-4946-4364-a499-bac7b8a5ec14 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.441700] env[62753]: DEBUG oslo_concurrency.lockutils [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.833s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.444068] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 9.892s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.444068] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.445583] env[62753]: DEBUG oslo_concurrency.lockutils [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 4.071s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1039.452645] env[62753]: DEBUG oslo_vmware.api [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1039.452645] env[62753]: value = "task-1332714" [ 1039.452645] env[62753]: _type = "Task" [ 1039.452645] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1039.460950] env[62753]: DEBUG oslo_vmware.api [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332714, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.463824] env[62753]: INFO nova.scheduler.client.report [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Deleted allocations for instance 9d7066e1-3721-42d2-8e80-91eacbbd8b80 [ 1039.467963] env[62753]: INFO nova.scheduler.client.report [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Deleted allocations for instance 2b90824a-5d49-4ec9-bcf6-4f868ca876cf [ 1039.626852] env[62753]: DEBUG nova.compute.manager [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1039.629951] env[62753]: INFO nova.compute.manager [-] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Took 1.41 seconds to deallocate network for instance. [ 1039.652918] env[62753]: DEBUG nova.virt.hardware [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1039.653411] env[62753]: DEBUG nova.virt.hardware [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1039.653411] env[62753]: DEBUG nova.virt.hardware [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1039.653558] env[62753]: DEBUG nova.virt.hardware [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1039.653658] env[62753]: DEBUG nova.virt.hardware [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1039.653814] env[62753]: DEBUG nova.virt.hardware [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1039.654036] env[62753]: DEBUG nova.virt.hardware [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1039.654208] env[62753]: DEBUG nova.virt.hardware [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1039.654427] env[62753]: DEBUG nova.virt.hardware [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1039.654605] env[62753]: DEBUG nova.virt.hardware [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1039.654783] env[62753]: DEBUG nova.virt.hardware [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1039.655678] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d35d81a3-8897-45d5-a837-3bef656c9758 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.664215] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3bdf346-91af-42b9-9ae3-8236f590b996 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1039.728156] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332712, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.763397] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332713, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1039.963048] env[62753]: DEBUG oslo_vmware.api [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332714, 'name': ReconfigVM_Task, 'duration_secs': 0.125594} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1039.963048] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284671', 'volume_id': 'e93861cc-b07a-4254-b33d-0814321eef34', 'name': 'volume-e93861cc-b07a-4254-b33d-0814321eef34', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '04fd0022-f20a-4217-9e47-5381635b17a6', 'attached_at': '', 'detached_at': '', 'volume_id': 'e93861cc-b07a-4254-b33d-0814321eef34', 'serial': 'e93861cc-b07a-4254-b33d-0814321eef34'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1039.976391] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ee516dbf-b601-4e93-9db6-e444529c0a31 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "2b90824a-5d49-4ec9-bcf6-4f868ca876cf" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.597s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1039.978231] env[62753]: DEBUG oslo_concurrency.lockutils [None req-67afd68b-015e-4d14-b1df-97087a36e4c8 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "9d7066e1-3721-42d2-8e80-91eacbbd8b80" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 14.234s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.117317] env[62753]: DEBUG nova.network.neutron [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Successfully updated port: 31af6b98-8cec-4501-9fd5-26eda99426d4 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1040.125816] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00bee189-8d38-4bea-a367-cc3a92823d57 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.134340] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e450bbc5-e4b0-406a-b01a-b428320f3d65 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.138325] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.168052] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d8780ca-1d4b-4431-bf27-394d4abcbb63 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.177442] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6c76cac6-a1c3-4add-b108-fe2bc056c5fe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.191740] env[62753]: DEBUG nova.compute.provider_tree [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1040.229258] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332712, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.263926] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332713, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.450377] env[62753]: DEBUG nova.compute.manager [req-c3d0564a-8560-4977-9e6f-67c10c01b70f req-5c5ce5d6-f126-4e60-8fcb-0a3cee13a8a9 service nova] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Received event network-vif-deleted-12ca481a-0cb3-465e-b04c-802ae3ce6438 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1040.450459] env[62753]: DEBUG nova.compute.manager [req-c3d0564a-8560-4977-9e6f-67c10c01b70f req-5c5ce5d6-f126-4e60-8fcb-0a3cee13a8a9 service nova] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Received event network-vif-plugged-31af6b98-8cec-4501-9fd5-26eda99426d4 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1040.450698] env[62753]: DEBUG oslo_concurrency.lockutils [req-c3d0564a-8560-4977-9e6f-67c10c01b70f req-5c5ce5d6-f126-4e60-8fcb-0a3cee13a8a9 service nova] Acquiring lock "0f81b85b-0550-46b9-b77c-9075a99cfaec-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1040.450931] env[62753]: DEBUG oslo_concurrency.lockutils [req-c3d0564a-8560-4977-9e6f-67c10c01b70f req-5c5ce5d6-f126-4e60-8fcb-0a3cee13a8a9 service nova] Lock "0f81b85b-0550-46b9-b77c-9075a99cfaec-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1040.451124] env[62753]: DEBUG oslo_concurrency.lockutils [req-c3d0564a-8560-4977-9e6f-67c10c01b70f req-5c5ce5d6-f126-4e60-8fcb-0a3cee13a8a9 service nova] Lock "0f81b85b-0550-46b9-b77c-9075a99cfaec-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1040.451298] env[62753]: DEBUG nova.compute.manager [req-c3d0564a-8560-4977-9e6f-67c10c01b70f req-5c5ce5d6-f126-4e60-8fcb-0a3cee13a8a9 service nova] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] No waiting events found dispatching network-vif-plugged-31af6b98-8cec-4501-9fd5-26eda99426d4 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1040.451470] env[62753]: WARNING nova.compute.manager [req-c3d0564a-8560-4977-9e6f-67c10c01b70f req-5c5ce5d6-f126-4e60-8fcb-0a3cee13a8a9 service nova] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Received unexpected event network-vif-plugged-31af6b98-8cec-4501-9fd5-26eda99426d4 for instance with vm_state building and task_state spawning. [ 1040.451635] env[62753]: DEBUG nova.compute.manager [req-c3d0564a-8560-4977-9e6f-67c10c01b70f req-5c5ce5d6-f126-4e60-8fcb-0a3cee13a8a9 service nova] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Received event network-changed-31af6b98-8cec-4501-9fd5-26eda99426d4 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1040.451791] env[62753]: DEBUG nova.compute.manager [req-c3d0564a-8560-4977-9e6f-67c10c01b70f req-5c5ce5d6-f126-4e60-8fcb-0a3cee13a8a9 service nova] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Refreshing instance network info cache due to event network-changed-31af6b98-8cec-4501-9fd5-26eda99426d4. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1040.451987] env[62753]: DEBUG oslo_concurrency.lockutils [req-c3d0564a-8560-4977-9e6f-67c10c01b70f req-5c5ce5d6-f126-4e60-8fcb-0a3cee13a8a9 service nova] Acquiring lock "refresh_cache-0f81b85b-0550-46b9-b77c-9075a99cfaec" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.452156] env[62753]: DEBUG oslo_concurrency.lockutils [req-c3d0564a-8560-4977-9e6f-67c10c01b70f req-5c5ce5d6-f126-4e60-8fcb-0a3cee13a8a9 service nova] Acquired lock "refresh_cache-0f81b85b-0550-46b9-b77c-9075a99cfaec" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1040.452318] env[62753]: DEBUG nova.network.neutron [req-c3d0564a-8560-4977-9e6f-67c10c01b70f req-5c5ce5d6-f126-4e60-8fcb-0a3cee13a8a9 service nova] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Refreshing network info cache for port 31af6b98-8cec-4501-9fd5-26eda99426d4 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1040.620449] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "refresh_cache-0f81b85b-0550-46b9-b77c-9075a99cfaec" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1040.695138] env[62753]: DEBUG nova.scheduler.client.report [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1040.730263] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332712, 'name': ReconfigVM_Task, 'duration_secs': 1.053253} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.730565] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Reconfigured VM instance instance-00000060 to attach disk [datastore2] 961ced74-10af-4443-857e-545cd890674e/961ced74-10af-4443-857e-545cd890674e.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1040.731231] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-5144dec3-4a24-47d9-9fce-02031f40ea6f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.738060] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1040.738060] env[62753]: value = "task-1332715" [ 1040.738060] env[62753]: _type = "Task" [ 1040.738060] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1040.747192] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332715, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1040.762798] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332713, 'name': PowerOnVM_Task, 'duration_secs': 1.188412} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1040.762990] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1040.763221] env[62753]: INFO nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Took 9.87 seconds to spawn the instance on the hypervisor. [ 1040.763415] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1040.764141] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f50f800-018c-4c82-9b39-bd4e6fd6b9b1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1040.993997] env[62753]: DEBUG nova.network.neutron [req-c3d0564a-8560-4977-9e6f-67c10c01b70f req-5c5ce5d6-f126-4e60-8fcb-0a3cee13a8a9 service nova] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1041.010072] env[62753]: DEBUG nova.objects.instance [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lazy-loading 'flavor' on Instance uuid 04fd0022-f20a-4217-9e47-5381635b17a6 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1041.075952] env[62753]: DEBUG oslo_concurrency.lockutils [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "bba5e3cb-c896-4ce7-ac7f-1cf2aba20465" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.076257] env[62753]: DEBUG oslo_concurrency.lockutils [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "bba5e3cb-c896-4ce7-ac7f-1cf2aba20465" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.076555] env[62753]: DEBUG oslo_concurrency.lockutils [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "bba5e3cb-c896-4ce7-ac7f-1cf2aba20465-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1041.076795] env[62753]: DEBUG oslo_concurrency.lockutils [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "bba5e3cb-c896-4ce7-ac7f-1cf2aba20465-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.077031] env[62753]: DEBUG oslo_concurrency.lockutils [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "bba5e3cb-c896-4ce7-ac7f-1cf2aba20465-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.079276] env[62753]: INFO nova.compute.manager [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Terminating instance [ 1041.081209] env[62753]: DEBUG nova.compute.manager [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1041.081410] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1041.082364] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-214a870e-465a-4fee-b90f-f52f5cfbf7c0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.090617] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1041.091161] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9e45e347-0b71-422d-85ec-75c6e7839110 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.095879] env[62753]: DEBUG nova.network.neutron [req-c3d0564a-8560-4977-9e6f-67c10c01b70f req-5c5ce5d6-f126-4e60-8fcb-0a3cee13a8a9 service nova] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1041.097349] env[62753]: DEBUG oslo_vmware.api [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1041.097349] env[62753]: value = "task-1332716" [ 1041.097349] env[62753]: _type = "Task" [ 1041.097349] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.108533] env[62753]: DEBUG oslo_vmware.api [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332716, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.248912] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332715, 'name': Rename_Task, 'duration_secs': 0.177107} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.249258] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1041.249534] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e412e4d7-01f1-4e12-9f5a-89cb1930e113 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.256996] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1041.256996] env[62753]: value = "task-1332717" [ 1041.256996] env[62753]: _type = "Task" [ 1041.256996] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.266857] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332717, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.282742] env[62753]: INFO nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Took 20.53 seconds to build instance. [ 1041.516584] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2a017a30-0c85-441f-9882-72d4c45eb5af tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "04fd0022-f20a-4217-9e47-5381635b17a6" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.304s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.602803] env[62753]: DEBUG oslo_concurrency.lockutils [req-c3d0564a-8560-4977-9e6f-67c10c01b70f req-5c5ce5d6-f126-4e60-8fcb-0a3cee13a8a9 service nova] Releasing lock "refresh_cache-0f81b85b-0550-46b9-b77c-9075a99cfaec" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1041.603293] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "refresh_cache-0f81b85b-0550-46b9-b77c-9075a99cfaec" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1041.603518] env[62753]: DEBUG nova.network.neutron [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1041.609657] env[62753]: DEBUG oslo_vmware.api [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332716, 'name': PowerOffVM_Task, 'duration_secs': 0.210419} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1041.609657] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1041.609657] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1041.610157] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-96a951c6-e6a1-467a-8c77-dfebedfd15a2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.679308] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1041.679624] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1041.679981] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Deleting the datastore file [datastore1] bba5e3cb-c896-4ce7-ac7f-1cf2aba20465 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1041.680682] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bd241e32-25ed-4e5c-abf7-deb34f40da80 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1041.687283] env[62753]: DEBUG oslo_vmware.api [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for the task: (returnval){ [ 1041.687283] env[62753]: value = "task-1332719" [ 1041.687283] env[62753]: _type = "Task" [ 1041.687283] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1041.695090] env[62753]: DEBUG oslo_vmware.api [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332719, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.705501] env[62753]: DEBUG oslo_concurrency.lockutils [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.260s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1041.705732] env[62753]: DEBUG nova.compute.manager [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Resized/migrated instance is powered off. Setting vm_state to 'stopped'. {{(pid=62753) _confirm_resize /opt/stack/nova/nova/compute/manager.py:4909}} [ 1041.708747] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.598s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1041.710119] env[62753]: INFO nova.compute.claims [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1041.768385] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332717, 'name': PowerOnVM_Task} progress is 90%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1041.785553] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "d7ba8be1-befb-4a0b-9c35-d86e7d586203" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.050s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.142494] env[62753]: DEBUG nova.network.neutron [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1042.198395] env[62753]: DEBUG oslo_vmware.api [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Task: {'id': task-1332719, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.2576} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.198912] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1042.198912] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1042.199248] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1042.199248] env[62753]: INFO nova.compute.manager [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1042.199455] env[62753]: DEBUG oslo.service.loopingcall [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.200019] env[62753]: DEBUG nova.compute.manager [-] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1042.200019] env[62753]: DEBUG nova.network.neutron [-] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1042.261739] env[62753]: INFO nova.scheduler.client.report [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Deleted allocation for migration 1c84581b-3dc0-4087-a0c2-fa39eb566788 [ 1042.269488] env[62753]: DEBUG oslo_vmware.api [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332717, 'name': PowerOnVM_Task, 'duration_secs': 0.562541} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1042.269952] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1042.270194] env[62753]: INFO nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Took 9.08 seconds to spawn the instance on the hypervisor. [ 1042.270335] env[62753]: DEBUG nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1042.271111] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f81ef31-4506-43e6-9289-32821b350cbf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.357403] env[62753]: DEBUG nova.network.neutron [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Updating instance_info_cache with network_info: [{"id": "31af6b98-8cec-4501-9fd5-26eda99426d4", "address": "fa:16:3e:4b:7c:4f", "network": {"id": "2823647f-d32d-4a1c-9e02-7577016a260e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1594498922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6c8a6bc946a469fa85cc3ab80d7333a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31af6b98-8c", "ovs_interfaceid": "31af6b98-8cec-4501-9fd5-26eda99426d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.509938] env[62753]: DEBUG nova.compute.manager [req-3dc0a1f6-eb42-42f6-b7e3-2c12bef04eea req-9b858ace-a661-4a92-b1d6-e74ba8f39152 service nova] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Received event network-vif-deleted-4832b8bd-7a18-4f8e-a05b-7421417d64ff {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1042.509938] env[62753]: INFO nova.compute.manager [req-3dc0a1f6-eb42-42f6-b7e3-2c12bef04eea req-9b858ace-a661-4a92-b1d6-e74ba8f39152 service nova] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Neutron deleted interface 4832b8bd-7a18-4f8e-a05b-7421417d64ff; detaching it from the instance and deleting it from the info cache [ 1042.509938] env[62753]: DEBUG nova.network.neutron [req-3dc0a1f6-eb42-42f6-b7e3-2c12bef04eea req-9b858ace-a661-4a92-b1d6-e74ba8f39152 service nova] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1042.675284] env[62753]: DEBUG oslo_concurrency.lockutils [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "01d9da0b-f4e5-474c-aab2-a0f4f0c99994" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1042.675747] env[62753]: DEBUG oslo_concurrency.lockutils [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "01d9da0b-f4e5-474c-aab2-a0f4f0c99994" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1042.770832] env[62753]: DEBUG oslo_concurrency.lockutils [None req-27f2c45e-0fa5-4c38-9cac-9c512788d50c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "c73b1ae0-dc98-47f7-babf-e96169384785" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 10.334s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1042.788659] env[62753]: INFO nova.compute.manager [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Took 22.02 seconds to build instance. [ 1042.859388] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "refresh_cache-0f81b85b-0550-46b9-b77c-9075a99cfaec" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1042.859705] env[62753]: DEBUG nova.compute.manager [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Instance network_info: |[{"id": "31af6b98-8cec-4501-9fd5-26eda99426d4", "address": "fa:16:3e:4b:7c:4f", "network": {"id": "2823647f-d32d-4a1c-9e02-7577016a260e", "bridge": "br-int", "label": "tempest-ServerDiskConfigTestJSON-1594498922-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "e6c8a6bc946a469fa85cc3ab80d7333a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "9d39252e-42ef-4252-98d3-62af5a0d109d", "external-id": "nsx-vlan-transportzone-190", "segmentation_id": 190, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap31af6b98-8c", "ovs_interfaceid": "31af6b98-8cec-4501-9fd5-26eda99426d4", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1042.860146] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:4b:7c:4f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '9d39252e-42ef-4252-98d3-62af5a0d109d', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '31af6b98-8cec-4501-9fd5-26eda99426d4', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1042.868539] env[62753]: DEBUG oslo.service.loopingcall [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1042.871458] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1042.872261] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-83ab5a6f-9b54-4918-b8a2-12835e967f7a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.892042] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-84a786c5-088a-4f07-b891-2ab652451df1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.900340] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c9d5271d-143a-446d-8fe9-da12bb56b964 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.905301] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1042.905301] env[62753]: value = "task-1332720" [ 1042.905301] env[62753]: _type = "Task" [ 1042.905301] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1042.937503] env[62753]: DEBUG nova.objects.instance [None req-e6afa000-a895-42fe-89e9-7111420e0a57 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lazy-loading 'flavor' on Instance uuid c73b1ae0-dc98-47f7-babf-e96169384785 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1042.941234] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8b15799-c8cf-402b-8987-57d08bdcb910 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.947934] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332720, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1042.954794] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1bc8921b-21d9-4be8-ac52-a1c2b435038f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1042.972871] env[62753]: DEBUG nova.compute.provider_tree [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1042.988420] env[62753]: DEBUG nova.network.neutron [-] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1043.012773] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-81c8b11d-34e4-4555-8a26-dec3cdedaa36 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.022621] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4120b72-322f-4667-beef-98f724ea3993 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.050436] env[62753]: DEBUG nova.compute.manager [req-3dc0a1f6-eb42-42f6-b7e3-2c12bef04eea req-9b858ace-a661-4a92-b1d6-e74ba8f39152 service nova] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Detach interface failed, port_id=4832b8bd-7a18-4f8e-a05b-7421417d64ff, reason: Instance bba5e3cb-c896-4ce7-ac7f-1cf2aba20465 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1043.179408] env[62753]: DEBUG nova.compute.manager [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1043.290821] env[62753]: DEBUG oslo_concurrency.lockutils [None req-50acfd62-53e1-4763-ae09-44a8d256d835 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "961ced74-10af-4443-857e-545cd890674e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.525s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.415747] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332720, 'name': CreateVM_Task, 'duration_secs': 0.394166} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.415980] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1043.416615] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.416808] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.417147] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1043.417401] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5966e045-0461-4bba-ab64-dcfc01d9f6a5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.421603] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 1043.421603] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523449b4-0bca-33fb-0e6e-d5f2cea46054" [ 1043.421603] env[62753]: _type = "Task" [ 1043.421603] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.428739] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523449b4-0bca-33fb-0e6e-d5f2cea46054, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.442958] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e6afa000-a895-42fe-89e9-7111420e0a57 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.443132] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e6afa000-a895-42fe-89e9-7111420e0a57 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.443337] env[62753]: DEBUG nova.network.neutron [None req-e6afa000-a895-42fe-89e9-7111420e0a57 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1043.443521] env[62753]: DEBUG nova.objects.instance [None req-e6afa000-a895-42fe-89e9-7111420e0a57 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lazy-loading 'info_cache' on Instance uuid c73b1ae0-dc98-47f7-babf-e96169384785 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.475874] env[62753]: DEBUG nova.scheduler.client.report [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1043.491123] env[62753]: INFO nova.compute.manager [-] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Took 1.29 seconds to deallocate network for instance. [ 1043.682815] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "ff50c5f4-ab27-4f17-948e-80e2dcc623eb" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.683119] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "ff50c5f4-ab27-4f17-948e-80e2dcc623eb" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.683339] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "ff50c5f4-ab27-4f17-948e-80e2dcc623eb-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.683531] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "ff50c5f4-ab27-4f17-948e-80e2dcc623eb-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.683704] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "ff50c5f4-ab27-4f17-948e-80e2dcc623eb-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.687839] env[62753]: INFO nova.compute.manager [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Terminating instance [ 1043.691521] env[62753]: DEBUG nova.compute.manager [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1043.691894] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1043.692598] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a76f8f6-c041-4485-ae2b-9c1e3184485d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.700545] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1043.700883] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-871df6fe-3e84-43ed-a4aa-fcc61ea2c23a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.704105] env[62753]: DEBUG oslo_concurrency.lockutils [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1043.706783] env[62753]: DEBUG oslo_vmware.api [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1043.706783] env[62753]: value = "task-1332721" [ 1043.706783] env[62753]: _type = "Task" [ 1043.706783] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.714836] env[62753]: DEBUG oslo_vmware.api [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332721, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.933060] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523449b4-0bca-33fb-0e6e-d5f2cea46054, 'name': SearchDatastore_Task, 'duration_secs': 0.021438} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1043.933060] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1043.933060] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1043.933060] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1043.933324] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1043.933324] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1043.933545] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-4a142a0e-a522-4a29-967a-ea8a5e617000 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.941364] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1043.941546] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1043.942250] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-33d0278a-4227-4111-b854-b70426e30494 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1043.946875] env[62753]: DEBUG nova.objects.base [None req-e6afa000-a895-42fe-89e9-7111420e0a57 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1043.949217] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 1043.949217] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5232bb04-135d-eb10-fcd2-47debd1a0b13" [ 1043.949217] env[62753]: _type = "Task" [ 1043.949217] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1043.957033] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5232bb04-135d-eb10-fcd2-47debd1a0b13, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1043.980661] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.272s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1043.981233] env[62753]: DEBUG nova.compute.manager [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1043.983997] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.808s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1043.984206] env[62753]: DEBUG nova.objects.instance [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lazy-loading 'resources' on Instance uuid 4d20e083-2959-453a-8875-47955bc02613 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1043.997144] env[62753]: DEBUG oslo_concurrency.lockutils [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1044.217042] env[62753]: DEBUG oslo_vmware.api [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332721, 'name': PowerOffVM_Task, 'duration_secs': 0.348593} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.217328] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1044.217502] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1044.217759] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9e003516-b9b1-43f1-bd6b-17e30bcda076 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.291202] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1044.291568] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1044.291634] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Deleting the datastore file [datastore2] ff50c5f4-ab27-4f17-948e-80e2dcc623eb {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1044.291872] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-6f76be48-87ad-4c8d-bebd-2bd3de4af3e5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.298577] env[62753]: DEBUG oslo_vmware.api [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1044.298577] env[62753]: value = "task-1332723" [ 1044.298577] env[62753]: _type = "Task" [ 1044.298577] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.306068] env[62753]: DEBUG oslo_vmware.api [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332723, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.465191] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5232bb04-135d-eb10-fcd2-47debd1a0b13, 'name': SearchDatastore_Task, 'duration_secs': 0.009528} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.466078] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-9e9cb06c-48aa-43ed-867d-c3ebfb977d54 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.471178] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 1044.471178] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]527b965c-8e8c-4630-50c1-55f41adc2063" [ 1044.471178] env[62753]: _type = "Task" [ 1044.471178] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1044.478673] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]527b965c-8e8c-4630-50c1-55f41adc2063, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1044.487171] env[62753]: DEBUG nova.compute.utils [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1044.493050] env[62753]: DEBUG nova.compute.manager [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1044.493227] env[62753]: DEBUG nova.network.neutron [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1044.555487] env[62753]: DEBUG nova.policy [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'ae52823f7c93454e8089261c3ca44321', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '1d52d6eaee934be5ab0e0003df1ce316', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1044.650181] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c57bb226-4f71-48a9-807c-1dc5fd323c53 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.659075] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66de48cd-2e12-46b0-bcda-600c02a41edb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.697117] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adcd04d7-f406-4a41-bc49-b9c09e970f33 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.706060] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4236034f-e078-433b-85ee-23afe222b1f8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.720532] env[62753]: DEBUG nova.compute.provider_tree [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1044.812392] env[62753]: DEBUG oslo_vmware.api [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332723, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.345229} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.812729] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1044.812943] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1044.813161] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1044.813345] env[62753]: INFO nova.compute.manager [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1044.813595] env[62753]: DEBUG oslo.service.loopingcall [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1044.813791] env[62753]: DEBUG nova.compute.manager [-] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1044.813887] env[62753]: DEBUG nova.network.neutron [-] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1044.817460] env[62753]: DEBUG nova.network.neutron [None req-e6afa000-a895-42fe-89e9-7111420e0a57 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating instance_info_cache with network_info: [{"id": "81591716-fb4a-427f-b3e9-a8b265fe1ce1", "address": "fa:16:3e:3e:ef:2c", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.7", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap81591716-fb", "ovs_interfaceid": "81591716-fb4a-427f-b3e9-a8b265fe1ce1", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1044.989026] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]527b965c-8e8c-4630-50c1-55f41adc2063, 'name': SearchDatastore_Task, 'duration_secs': 0.017251} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1044.989794] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1044.989794] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 0f81b85b-0550-46b9-b77c-9075a99cfaec/0f81b85b-0550-46b9-b77c-9075a99cfaec.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1044.990010] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-39fe806c-8172-4c58-9ec7-c08c122d6cbf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1044.993923] env[62753]: DEBUG nova.compute.manager [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1044.998410] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 1044.998410] env[62753]: value = "task-1332724" [ 1044.998410] env[62753]: _type = "Task" [ 1044.998410] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.008114] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332724, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.018556] env[62753]: DEBUG nova.network.neutron [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Successfully created port: 93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1045.220567] env[62753]: DEBUG nova.compute.manager [req-7f382a28-ab95-416f-9954-a16f2d8a41c6 req-085b25f8-f872-4805-baab-0d825925beb3 service nova] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Received event network-vif-deleted-86406a4f-4de3-4c2d-b326-ced3cbdf893b {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1045.220808] env[62753]: INFO nova.compute.manager [req-7f382a28-ab95-416f-9954-a16f2d8a41c6 req-085b25f8-f872-4805-baab-0d825925beb3 service nova] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Neutron deleted interface 86406a4f-4de3-4c2d-b326-ced3cbdf893b; detaching it from the instance and deleting it from the info cache [ 1045.221076] env[62753]: DEBUG nova.network.neutron [req-7f382a28-ab95-416f-9954-a16f2d8a41c6 req-085b25f8-f872-4805-baab-0d825925beb3 service nova] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.223157] env[62753]: DEBUG nova.scheduler.client.report [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1045.320116] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e6afa000-a895-42fe-89e9-7111420e0a57 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "refresh_cache-c73b1ae0-dc98-47f7-babf-e96169384785" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1045.513282] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332724, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.700233] env[62753]: DEBUG nova.network.neutron [-] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1045.723097] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cb86573e-676f-4bfd-9a50-7cce43791e81 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.728371] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.744s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1045.731382] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.595s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1045.731625] env[62753]: DEBUG nova.objects.instance [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lazy-loading 'resources' on Instance uuid b1e4c8f9-471d-4cca-9a12-56191767a5b1 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1045.735259] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3f9b3e0-b46d-4b4f-bb5a-821b1c99d725 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.751093] env[62753]: INFO nova.scheduler.client.report [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Deleted allocations for instance 4d20e083-2959-453a-8875-47955bc02613 [ 1045.765229] env[62753]: DEBUG nova.compute.manager [req-7f382a28-ab95-416f-9954-a16f2d8a41c6 req-085b25f8-f872-4805-baab-0d825925beb3 service nova] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Detach interface failed, port_id=86406a4f-4de3-4c2d-b326-ced3cbdf893b, reason: Instance ff50c5f4-ab27-4f17-948e-80e2dcc623eb could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1045.823337] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6afa000-a895-42fe-89e9-7111420e0a57 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1045.823817] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-1b6ea511-edef-4005-b0ea-bcbc9736e8ac {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.832483] env[62753]: DEBUG oslo_vmware.api [None req-e6afa000-a895-42fe-89e9-7111420e0a57 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1045.832483] env[62753]: value = "task-1332725" [ 1045.832483] env[62753]: _type = "Task" [ 1045.832483] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1045.842078] env[62753]: DEBUG oslo_vmware.api [None req-e6afa000-a895-42fe-89e9-7111420e0a57 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332725, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1045.889862] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-91ceddde-3242-472e-a916-ae78a315c24e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.897516] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30b33535-5771-4c24-a83b-e481b25dc1da {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.926491] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f970154d-e81a-45e9-8388-beb551f94a1f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.934082] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a997fdb6-aa81-4fb2-98b8-d83451663f3d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1045.948167] env[62753]: DEBUG nova.compute.provider_tree [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1046.008396] env[62753]: DEBUG nova.compute.manager [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1046.013539] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332724, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.567267} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.014000] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 0f81b85b-0550-46b9-b77c-9075a99cfaec/0f81b85b-0550-46b9-b77c-9075a99cfaec.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1046.014000] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1046.014659] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-64bde833-d342-4e46-a1dc-0a5985dc9ab1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.020971] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 1046.020971] env[62753]: value = "task-1332726" [ 1046.020971] env[62753]: _type = "Task" [ 1046.020971] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.029303] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332726, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.042419] env[62753]: DEBUG nova.virt.hardware [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1046.042728] env[62753]: DEBUG nova.virt.hardware [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1046.042979] env[62753]: DEBUG nova.virt.hardware [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1046.043255] env[62753]: DEBUG nova.virt.hardware [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1046.043524] env[62753]: DEBUG nova.virt.hardware [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1046.043737] env[62753]: DEBUG nova.virt.hardware [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1046.044066] env[62753]: DEBUG nova.virt.hardware [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1046.044302] env[62753]: DEBUG nova.virt.hardware [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1046.044543] env[62753]: DEBUG nova.virt.hardware [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1046.044739] env[62753]: DEBUG nova.virt.hardware [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1046.044928] env[62753]: DEBUG nova.virt.hardware [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1046.045810] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b2ff368-3af2-4620-83b2-eddc51ed96dc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.053566] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6fab6ea1-fe09-4b81-bcf4-b72cebaf43d6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.203415] env[62753]: INFO nova.compute.manager [-] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Took 1.39 seconds to deallocate network for instance. [ 1046.258892] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d8020286-0fc5-46d3-b941-164a22958cd7 tempest-ServersNegativeTestJSON-1018058270 tempest-ServersNegativeTestJSON-1018058270-project-member] Lock "4d20e083-2959-453a-8875-47955bc02613" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 13.771s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.345583] env[62753]: DEBUG oslo_vmware.api [None req-e6afa000-a895-42fe-89e9-7111420e0a57 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332725, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.451524] env[62753]: DEBUG nova.scheduler.client.report [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1046.531648] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332726, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.072255} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.531927] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1046.532746] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47490668-92b8-46a5-a7b0-e4deedfbdc3a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.556290] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Reconfiguring VM instance instance-00000061 to attach disk [datastore1] 0f81b85b-0550-46b9-b77c-9075a99cfaec/0f81b85b-0550-46b9-b77c-9075a99cfaec.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1046.556290] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b85a5218-fc2e-47b8-a707-60f3f28f9b5c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.576604] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 1046.576604] env[62753]: value = "task-1332727" [ 1046.576604] env[62753]: _type = "Task" [ 1046.576604] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1046.585718] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332727, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1046.586588] env[62753]: DEBUG nova.network.neutron [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Successfully updated port: 93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1046.710076] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1046.845863] env[62753]: DEBUG oslo_vmware.api [None req-e6afa000-a895-42fe-89e9-7111420e0a57 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332725, 'name': PowerOnVM_Task, 'duration_secs': 0.889688} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1046.846233] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6afa000-a895-42fe-89e9-7111420e0a57 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1046.846470] env[62753]: DEBUG nova.compute.manager [None req-e6afa000-a895-42fe-89e9-7111420e0a57 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1046.847402] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e070d5b1-6ce7-469b-9873-f34cd821077d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1046.956794] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.225s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1046.959987] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.821s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1046.959987] env[62753]: DEBUG nova.objects.instance [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lazy-loading 'resources' on Instance uuid c3679c66-2763-4a04-8d14-e103bd952798 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1046.988018] env[62753]: INFO nova.scheduler.client.report [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Deleted allocations for instance b1e4c8f9-471d-4cca-9a12-56191767a5b1 [ 1047.087726] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332727, 'name': ReconfigVM_Task, 'duration_secs': 0.333303} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.088039] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Reconfigured VM instance instance-00000061 to attach disk [datastore1] 0f81b85b-0550-46b9-b77c-9075a99cfaec/0f81b85b-0550-46b9-b77c-9075a99cfaec.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1047.088955] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "refresh_cache-eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.089135] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "refresh_cache-eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1047.089229] env[62753]: DEBUG nova.network.neutron [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1047.090286] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-cc6ebaad-12f4-4777-9254-265a59e4d666 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.101018] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 1047.101018] env[62753]: value = "task-1332728" [ 1047.101018] env[62753]: _type = "Task" [ 1047.101018] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.108603] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332728, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.250256] env[62753]: DEBUG nova.compute.manager [req-0b637ade-7edd-403d-b372-1d42c8cc7bc9 req-42b1a1a0-2704-4035-bdb3-160fe2e6031e service nova] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Received event network-vif-plugged-93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1047.250435] env[62753]: DEBUG oslo_concurrency.lockutils [req-0b637ade-7edd-403d-b372-1d42c8cc7bc9 req-42b1a1a0-2704-4035-bdb3-160fe2e6031e service nova] Acquiring lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1047.250540] env[62753]: DEBUG oslo_concurrency.lockutils [req-0b637ade-7edd-403d-b372-1d42c8cc7bc9 req-42b1a1a0-2704-4035-bdb3-160fe2e6031e service nova] Lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1047.250705] env[62753]: DEBUG oslo_concurrency.lockutils [req-0b637ade-7edd-403d-b372-1d42c8cc7bc9 req-42b1a1a0-2704-4035-bdb3-160fe2e6031e service nova] Lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.251660] env[62753]: DEBUG nova.compute.manager [req-0b637ade-7edd-403d-b372-1d42c8cc7bc9 req-42b1a1a0-2704-4035-bdb3-160fe2e6031e service nova] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] No waiting events found dispatching network-vif-plugged-93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1047.251860] env[62753]: WARNING nova.compute.manager [req-0b637ade-7edd-403d-b372-1d42c8cc7bc9 req-42b1a1a0-2704-4035-bdb3-160fe2e6031e service nova] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Received unexpected event network-vif-plugged-93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9 for instance with vm_state building and task_state spawning. [ 1047.252231] env[62753]: DEBUG nova.compute.manager [req-0b637ade-7edd-403d-b372-1d42c8cc7bc9 req-42b1a1a0-2704-4035-bdb3-160fe2e6031e service nova] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Received event network-changed-93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1047.252231] env[62753]: DEBUG nova.compute.manager [req-0b637ade-7edd-403d-b372-1d42c8cc7bc9 req-42b1a1a0-2704-4035-bdb3-160fe2e6031e service nova] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Refreshing instance network info cache due to event network-changed-93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1047.252430] env[62753]: DEBUG oslo_concurrency.lockutils [req-0b637ade-7edd-403d-b372-1d42c8cc7bc9 req-42b1a1a0-2704-4035-bdb3-160fe2e6031e service nova] Acquiring lock "refresh_cache-eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1047.494939] env[62753]: DEBUG oslo_concurrency.lockutils [None req-04a47ff2-e9a5-4587-8e81-4209fe3e9b9b tempest-ServersTestJSON-646230718 tempest-ServersTestJSON-646230718-project-member] Lock "b1e4c8f9-471d-4cca-9a12-56191767a5b1" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.260s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1047.610635] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332728, 'name': Rename_Task, 'duration_secs': 0.149654} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1047.610971] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1047.614591] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-661be82e-c3ef-4c42-9570-d49911e17650 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.621916] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 1047.621916] env[62753]: value = "task-1332729" [ 1047.621916] env[62753]: _type = "Task" [ 1047.621916] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1047.625849] env[62753]: DEBUG nova.network.neutron [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1047.630034] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-868f9786-32e8-4763-b0d2-8d455a13c0f7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.635859] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332729, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1047.641289] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f2d35774-18a2-492f-b976-b7cf68c7de8c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.679095] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4131db33-713c-44a9-a82c-c3ce58074c1d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.688248] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d7bcde58-a148-43f2-aa23-62c9001f8565 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1047.704351] env[62753]: DEBUG nova.compute.provider_tree [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1047.812862] env[62753]: DEBUG nova.network.neutron [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Updating instance_info_cache with network_info: [{"id": "93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9", "address": "fa:16:3e:90:68:e2", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93d26d06-fe", "ovs_interfaceid": "93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.132737] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332729, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.208028] env[62753]: DEBUG nova.scheduler.client.report [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1048.319030] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "refresh_cache-eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1048.319030] env[62753]: DEBUG nova.compute.manager [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Instance network_info: |[{"id": "93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9", "address": "fa:16:3e:90:68:e2", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93d26d06-fe", "ovs_interfaceid": "93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1048.319030] env[62753]: DEBUG oslo_concurrency.lockutils [req-0b637ade-7edd-403d-b372-1d42c8cc7bc9 req-42b1a1a0-2704-4035-bdb3-160fe2e6031e service nova] Acquired lock "refresh_cache-eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.319030] env[62753]: DEBUG nova.network.neutron [req-0b637ade-7edd-403d-b372-1d42c8cc7bc9 req-42b1a1a0-2704-4035-bdb3-160fe2e6031e service nova] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Refreshing network info cache for port 93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1048.319030] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:68:e2', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a10c88d7-d13f-44fd-acee-7a734eb5f56a', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1048.335650] env[62753]: DEBUG oslo.service.loopingcall [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1048.339344] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1048.340059] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d013646f-2a29-4724-813d-82eddcc7cd0d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.368735] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1048.368735] env[62753]: value = "task-1332730" [ 1048.368735] env[62753]: _type = "Task" [ 1048.368735] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.379924] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332730, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.562920] env[62753]: DEBUG nova.network.neutron [req-0b637ade-7edd-403d-b372-1d42c8cc7bc9 req-42b1a1a0-2704-4035-bdb3-160fe2e6031e service nova] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Updated VIF entry in instance network info cache for port 93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1048.563367] env[62753]: DEBUG nova.network.neutron [req-0b637ade-7edd-403d-b372-1d42c8cc7bc9 req-42b1a1a0-2704-4035-bdb3-160fe2e6031e service nova] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Updating instance_info_cache with network_info: [{"id": "93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9", "address": "fa:16:3e:90:68:e2", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93d26d06-fe", "ovs_interfaceid": "93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1048.632371] env[62753]: DEBUG oslo_vmware.api [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332729, 'name': PowerOnVM_Task, 'duration_secs': 0.757808} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.632682] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1048.633160] env[62753]: INFO nova.compute.manager [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Took 9.01 seconds to spawn the instance on the hypervisor. [ 1048.633160] env[62753]: DEBUG nova.compute.manager [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1048.633875] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1797930-51f6-43d5-b033-7bfed6f6c9e6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.714435] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.755s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.716889] env[62753]: DEBUG oslo_concurrency.lockutils [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.013s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.718631] env[62753]: INFO nova.compute.claims [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1048.749160] env[62753]: INFO nova.scheduler.client.report [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Deleted allocations for instance c3679c66-2763-4a04-8d14-e103bd952798 [ 1048.866052] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "c73b1ae0-dc98-47f7-babf-e96169384785" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.866342] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "c73b1ae0-dc98-47f7-babf-e96169384785" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.866576] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "c73b1ae0-dc98-47f7-babf-e96169384785-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1048.866796] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "c73b1ae0-dc98-47f7-babf-e96169384785-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1048.867098] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "c73b1ae0-dc98-47f7-babf-e96169384785-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1048.869214] env[62753]: INFO nova.compute.manager [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Terminating instance [ 1048.874433] env[62753]: DEBUG nova.compute.manager [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1048.874689] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1048.875627] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d1d16236-ede8-4be6-822b-1838ec258b8e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.883558] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332730, 'name': CreateVM_Task, 'duration_secs': 0.504122} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1048.885693] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1048.885986] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1048.886705] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1048.886879] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1048.887193] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1048.887422] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bd54e799-4670-4223-80fd-b413dfe1997c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.888910] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7be88b81-9ed1-4482-8ef3-8bade8474056 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1048.893247] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1048.893247] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]528f070c-2524-5fd0-5800-3a29c076aba5" [ 1048.893247] env[62753]: _type = "Task" [ 1048.893247] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.899011] env[62753]: DEBUG oslo_vmware.api [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1048.899011] env[62753]: value = "task-1332731" [ 1048.899011] env[62753]: _type = "Task" [ 1048.899011] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1048.904932] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]528f070c-2524-5fd0-5800-3a29c076aba5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1048.909674] env[62753]: DEBUG oslo_vmware.api [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332731, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.065900] env[62753]: DEBUG oslo_concurrency.lockutils [req-0b637ade-7edd-403d-b372-1d42c8cc7bc9 req-42b1a1a0-2704-4035-bdb3-160fe2e6031e service nova] Releasing lock "refresh_cache-eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.152810] env[62753]: INFO nova.compute.manager [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Took 21.25 seconds to build instance. [ 1049.258149] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a3e2bcda-880c-4878-bba6-bb50e177abe5 tempest-AttachInterfacesTestJSON-816866237 tempest-AttachInterfacesTestJSON-816866237-project-member] Lock "c3679c66-2763-4a04-8d14-e103bd952798" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.852s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.407291] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]528f070c-2524-5fd0-5800-3a29c076aba5, 'name': SearchDatastore_Task, 'duration_secs': 0.010094} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.407606] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.407843] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1049.408092] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1049.408247] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1049.408432] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1049.409070] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-3f79a6e8-8734-4580-ac54-a19912f581e6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.413763] env[62753]: DEBUG oslo_vmware.api [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332731, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.421828] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1049.421828] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1049.422013] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a2a20030-d7fd-4061-bfa3-dcdbe4c48608 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.427790] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1049.427790] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5256225c-6381-3c6c-01c6-7132f36bdf4b" [ 1049.427790] env[62753]: _type = "Task" [ 1049.427790] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.435039] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5256225c-6381-3c6c-01c6-7132f36bdf4b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.655085] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d3313e24-d554-405a-bd8f-c849b6478e9d tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "0f81b85b-0550-46b9-b77c-9075a99cfaec" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 22.755s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1049.913538] env[62753]: DEBUG oslo_vmware.api [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332731, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1049.939891] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5256225c-6381-3c6c-01c6-7132f36bdf4b, 'name': SearchDatastore_Task, 'duration_secs': 0.031921} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.942984] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95abbb65-edfd-4f43-850a-8d8c4d36871e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.946041] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5a4882d3-aadf-4b4a-8558-0f305a3219c6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.954833] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21253a01-9dcb-4455-ae4d-6d79409998a9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.958258] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1049.958258] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ea5a30-b362-059b-16dc-d941f8ee29d6" [ 1049.958258] env[62753]: _type = "Task" [ 1049.958258] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1049.990379] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db08f058-904d-49f8-bbb1-3e575aefaf67 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1049.997022] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ea5a30-b362-059b-16dc-d941f8ee29d6, 'name': SearchDatastore_Task, 'duration_secs': 0.014968} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1049.997856] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1049.998214] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] eb86a978-27fa-41b2-a7e5-4b3a8ba0a152/eb86a978-27fa-41b2-a7e5-4b3a8ba0a152.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1049.998556] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ba5687e9-966b-44bd-b88f-5c1c7cf2d3ea {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.004882] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1ad6ad3-79b4-46fd-aade-4c7717bdd1e7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.010176] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1050.010176] env[62753]: value = "task-1332732" [ 1050.010176] env[62753]: _type = "Task" [ 1050.010176] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.023594] env[62753]: DEBUG nova.compute.provider_tree [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1050.030485] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332732, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.411703] env[62753]: DEBUG oslo_vmware.api [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332731, 'name': PowerOffVM_Task, 'duration_secs': 1.105226} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1050.411990] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1050.412182] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1050.412462] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-9c449946-1bce-4fda-9461-ed50e71dad4a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.492834] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1050.493267] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1050.493606] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Deleting the datastore file [datastore1] c73b1ae0-dc98-47f7-babf-e96169384785 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1050.494063] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-0d6d4d08-58bd-4cc8-b560-2a477342603a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1050.501882] env[62753]: DEBUG oslo_vmware.api [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1050.501882] env[62753]: value = "task-1332734" [ 1050.501882] env[62753]: _type = "Task" [ 1050.501882] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1050.513781] env[62753]: DEBUG oslo_vmware.api [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332734, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.523268] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332732, 'name': CopyVirtualDisk_Task} progress is 4%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1050.528182] env[62753]: DEBUG nova.scheduler.client.report [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1051.012840] env[62753]: DEBUG oslo_vmware.api [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332734, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.024246] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332732, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.906} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.024724] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] eb86a978-27fa-41b2-a7e5-4b3a8ba0a152/eb86a978-27fa-41b2-a7e5-4b3a8ba0a152.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1051.025041] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1051.025441] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-0898f7f3-6c39-406d-b719-67bb119c3723 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.032192] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1051.032192] env[62753]: value = "task-1332735" [ 1051.032192] env[62753]: _type = "Task" [ 1051.032192] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.032667] env[62753]: DEBUG oslo_concurrency.lockutils [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.316s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.033217] env[62753]: DEBUG nova.compute.manager [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1051.041635] env[62753]: DEBUG oslo_concurrency.lockutils [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.045s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.042055] env[62753]: DEBUG nova.objects.instance [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lazy-loading 'resources' on Instance uuid bba5e3cb-c896-4ce7-ac7f-1cf2aba20465 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1051.049159] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332735, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.221429] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "0f81b85b-0550-46b9-b77c-9075a99cfaec" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.221725] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "0f81b85b-0550-46b9-b77c-9075a99cfaec" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.221986] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "0f81b85b-0550-46b9-b77c-9075a99cfaec-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1051.222212] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "0f81b85b-0550-46b9-b77c-9075a99cfaec-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1051.222416] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "0f81b85b-0550-46b9-b77c-9075a99cfaec-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1051.224808] env[62753]: INFO nova.compute.manager [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Terminating instance [ 1051.226821] env[62753]: DEBUG nova.compute.manager [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1051.227105] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1051.227955] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ea92158-98cc-40fd-a79e-2db98e45569b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.235669] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1051.238837] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-acb9718f-b289-4083-884d-3ad4e67425c0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.243160] env[62753]: DEBUG oslo_vmware.api [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 1051.243160] env[62753]: value = "task-1332736" [ 1051.243160] env[62753]: _type = "Task" [ 1051.243160] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.251583] env[62753]: DEBUG oslo_vmware.api [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332736, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.512952] env[62753]: DEBUG oslo_vmware.api [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332734, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.554444} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.513279] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1051.513492] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1051.513675] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1051.513852] env[62753]: INFO nova.compute.manager [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Took 2.64 seconds to destroy the instance on the hypervisor. [ 1051.514137] env[62753]: DEBUG oslo.service.loopingcall [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1051.514349] env[62753]: DEBUG nova.compute.manager [-] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1051.514461] env[62753]: DEBUG nova.network.neutron [-] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1051.542880] env[62753]: DEBUG nova.compute.utils [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1051.544518] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332735, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064792} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.544622] env[62753]: DEBUG nova.compute.manager [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1051.544737] env[62753]: DEBUG nova.network.neutron [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1051.548933] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1051.550592] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6be0e40a-ee2f-4c78-a169-659f32a93d2b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.574050] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] eb86a978-27fa-41b2-a7e5-4b3a8ba0a152/eb86a978-27fa-41b2-a7e5-4b3a8ba0a152.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1051.577297] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5e3629e0-f595-4000-8c78-c406b7f0b72e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.598546] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1051.598546] env[62753]: value = "task-1332737" [ 1051.598546] env[62753]: _type = "Task" [ 1051.598546] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.609932] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332737, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1051.620468] env[62753]: DEBUG nova.policy [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '592d1d366cc4461299dbc28cee63e5b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8925b84dcf9a47fbaf2eb8044b3850fa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1051.719885] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4606ad0-aad6-44eb-a141-9c51c4bce8ed {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.727722] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-02481cb8-b293-453d-a465-7236bd35fde4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.762529] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcbe4632-f20f-4a61-8e81-dcb5cbaf7fda {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.772627] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be09344-35a1-4750-97e1-aa19ba993759 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.776428] env[62753]: DEBUG oslo_vmware.api [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332736, 'name': PowerOffVM_Task, 'duration_secs': 0.410924} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1051.776691] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1051.776888] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1051.777449] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-179c402f-a763-412a-86fb-66462c1137cd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.787909] env[62753]: DEBUG nova.compute.provider_tree [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1051.857494] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1051.857737] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1051.858006] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Deleting the datastore file [datastore1] 0f81b85b-0550-46b9-b77c-9075a99cfaec {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1051.858306] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d61f6c2d-22b7-44ee-bdcb-dd251fdababc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1051.864544] env[62753]: DEBUG oslo_vmware.api [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for the task: (returnval){ [ 1051.864544] env[62753]: value = "task-1332739" [ 1051.864544] env[62753]: _type = "Task" [ 1051.864544] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1051.872878] env[62753]: DEBUG oslo_vmware.api [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332739, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.050941] env[62753]: DEBUG nova.compute.manager [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1052.077580] env[62753]: DEBUG nova.network.neutron [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Successfully created port: d3934283-dc65-4a50-8b00-c466bb372792 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1052.110504] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332737, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.290748] env[62753]: DEBUG nova.scheduler.client.report [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1052.301978] env[62753]: DEBUG nova.compute.manager [req-5c7e6d8f-31b6-4351-ae53-80150cca8c2d req-45528419-a9b4-458e-9265-c9c9156d55d8 service nova] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Received event network-vif-deleted-81591716-fb4a-427f-b3e9-a8b265fe1ce1 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1052.302199] env[62753]: INFO nova.compute.manager [req-5c7e6d8f-31b6-4351-ae53-80150cca8c2d req-45528419-a9b4-458e-9265-c9c9156d55d8 service nova] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Neutron deleted interface 81591716-fb4a-427f-b3e9-a8b265fe1ce1; detaching it from the instance and deleting it from the info cache [ 1052.302382] env[62753]: DEBUG nova.network.neutron [req-5c7e6d8f-31b6-4351-ae53-80150cca8c2d req-45528419-a9b4-458e-9265-c9c9156d55d8 service nova] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.375476] env[62753]: DEBUG oslo_vmware.api [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Task: {'id': task-1332739, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.314308} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.375769] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1052.375960] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1052.376174] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1052.376353] env[62753]: INFO nova.compute.manager [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Took 1.15 seconds to destroy the instance on the hypervisor. [ 1052.376596] env[62753]: DEBUG oslo.service.loopingcall [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1052.376803] env[62753]: DEBUG nova.compute.manager [-] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1052.376902] env[62753]: DEBUG nova.network.neutron [-] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1052.613496] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332737, 'name': ReconfigVM_Task, 'duration_secs': 0.632591} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1052.613496] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Reconfigured VM instance instance-00000062 to attach disk [datastore1] eb86a978-27fa-41b2-a7e5-4b3a8ba0a152/eb86a978-27fa-41b2-a7e5-4b3a8ba0a152.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1052.613496] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b0011c2e-29ee-4e0d-9e29-9a63ad8711cd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.618393] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1052.618393] env[62753]: value = "task-1332740" [ 1052.618393] env[62753]: _type = "Task" [ 1052.618393] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1052.630415] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332740, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1052.773050] env[62753]: DEBUG nova.network.neutron [-] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1052.795698] env[62753]: DEBUG oslo_concurrency.lockutils [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.754s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1052.798032] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 6.088s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1052.798096] env[62753]: DEBUG nova.objects.instance [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lazy-loading 'resources' on Instance uuid ff50c5f4-ab27-4f17-948e-80e2dcc623eb {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1052.805454] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-41104d5c-7cd9-4dc8-b977-2754abaf7f98 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.817575] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-233ae56b-3d5b-4c68-b797-1b2541a24390 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1052.829181] env[62753]: INFO nova.scheduler.client.report [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Deleted allocations for instance bba5e3cb-c896-4ce7-ac7f-1cf2aba20465 [ 1052.848282] env[62753]: DEBUG nova.compute.manager [req-5c7e6d8f-31b6-4351-ae53-80150cca8c2d req-45528419-a9b4-458e-9265-c9c9156d55d8 service nova] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Detach interface failed, port_id=81591716-fb4a-427f-b3e9-a8b265fe1ce1, reason: Instance c73b1ae0-dc98-47f7-babf-e96169384785 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1053.062211] env[62753]: DEBUG nova.compute.manager [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1053.083105] env[62753]: DEBUG nova.virt.hardware [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1053.083105] env[62753]: DEBUG nova.virt.hardware [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1053.083105] env[62753]: DEBUG nova.virt.hardware [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1053.083105] env[62753]: DEBUG nova.virt.hardware [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1053.083295] env[62753]: DEBUG nova.virt.hardware [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1053.083369] env[62753]: DEBUG nova.virt.hardware [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1053.083593] env[62753]: DEBUG nova.virt.hardware [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1053.083751] env[62753]: DEBUG nova.virt.hardware [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1053.083923] env[62753]: DEBUG nova.virt.hardware [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1053.084125] env[62753]: DEBUG nova.virt.hardware [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1053.084791] env[62753]: DEBUG nova.virt.hardware [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1053.085252] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8f956780-cb37-465a-9a52-a1feeb373cf6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.093018] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-025bf478-c094-41e5-8721-93f39259ddae {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.128801] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332740, 'name': Rename_Task, 'duration_secs': 0.196368} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1053.129147] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1053.129400] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-949b758f-ea96-4a1b-b743-4548d849a63e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.135771] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1053.135771] env[62753]: value = "task-1332741" [ 1053.135771] env[62753]: _type = "Task" [ 1053.135771] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1053.144564] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332741, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.166133] env[62753]: DEBUG nova.network.neutron [-] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1053.275579] env[62753]: INFO nova.compute.manager [-] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Took 1.76 seconds to deallocate network for instance. [ 1053.337058] env[62753]: DEBUG oslo_concurrency.lockutils [None req-755c163c-960c-4735-bc08-35f4e84d2d23 tempest-ServerRescueTestJSON-799009928 tempest-ServerRescueTestJSON-799009928-project-member] Lock "bba5e3cb-c896-4ce7-ac7f-1cf2aba20465" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.261s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1053.409806] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29b43276-860f-4308-950a-0b705e97ec8d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.417534] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f035135-63df-4f14-92ee-4820a6db6da8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.449339] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-109229c9-7694-4771-8052-4db8aa52ca3d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.456848] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c1cc4f71-8ce2-4996-958b-31ca6cacc1a8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1053.470823] env[62753]: DEBUG nova.compute.provider_tree [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1053.582273] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "6e417227-895d-4576-b025-7d4a0aafa379" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.582530] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "6e417227-895d-4576-b025-7d4a0aafa379" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.646206] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332741, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1053.669660] env[62753]: INFO nova.compute.manager [-] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Took 1.29 seconds to deallocate network for instance. [ 1053.778812] env[62753]: DEBUG nova.network.neutron [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Successfully updated port: d3934283-dc65-4a50-8b00-c466bb372792 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1053.781695] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.908061] env[62753]: DEBUG oslo_concurrency.lockutils [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "32563e1f-398f-4fc9-866f-d85d9f3f27a7" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1053.908294] env[62753]: DEBUG oslo_concurrency.lockutils [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "32563e1f-398f-4fc9-866f-d85d9f3f27a7" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1053.974518] env[62753]: DEBUG nova.scheduler.client.report [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1054.088765] env[62753]: DEBUG nova.compute.manager [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1054.146906] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332741, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.176657] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.221664] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.222137] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.285149] env[62753]: DEBUG oslo_concurrency.lockutils [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "refresh_cache-01d9da0b-f4e5-474c-aab2-a0f4f0c99994" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.285149] env[62753]: DEBUG oslo_concurrency.lockutils [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquired lock "refresh_cache-01d9da0b-f4e5-474c-aab2-a0f4f0c99994" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1054.285149] env[62753]: DEBUG nova.network.neutron [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1054.343150] env[62753]: DEBUG nova.compute.manager [req-75ba0c6f-bc63-4d4c-ac50-338f73660bb2 req-e8d4a152-553c-4826-9dd8-c162ca665390 service nova] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Received event network-vif-deleted-31af6b98-8cec-4501-9fd5-26eda99426d4 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1054.343150] env[62753]: DEBUG nova.compute.manager [req-75ba0c6f-bc63-4d4c-ac50-338f73660bb2 req-e8d4a152-553c-4826-9dd8-c162ca665390 service nova] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Received event network-vif-plugged-d3934283-dc65-4a50-8b00-c466bb372792 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1054.343150] env[62753]: DEBUG oslo_concurrency.lockutils [req-75ba0c6f-bc63-4d4c-ac50-338f73660bb2 req-e8d4a152-553c-4826-9dd8-c162ca665390 service nova] Acquiring lock "01d9da0b-f4e5-474c-aab2-a0f4f0c99994-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.343150] env[62753]: DEBUG oslo_concurrency.lockutils [req-75ba0c6f-bc63-4d4c-ac50-338f73660bb2 req-e8d4a152-553c-4826-9dd8-c162ca665390 service nova] Lock "01d9da0b-f4e5-474c-aab2-a0f4f0c99994-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.343150] env[62753]: DEBUG oslo_concurrency.lockutils [req-75ba0c6f-bc63-4d4c-ac50-338f73660bb2 req-e8d4a152-553c-4826-9dd8-c162ca665390 service nova] Lock "01d9da0b-f4e5-474c-aab2-a0f4f0c99994-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.344269] env[62753]: DEBUG nova.compute.manager [req-75ba0c6f-bc63-4d4c-ac50-338f73660bb2 req-e8d4a152-553c-4826-9dd8-c162ca665390 service nova] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] No waiting events found dispatching network-vif-plugged-d3934283-dc65-4a50-8b00-c466bb372792 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1054.344636] env[62753]: WARNING nova.compute.manager [req-75ba0c6f-bc63-4d4c-ac50-338f73660bb2 req-e8d4a152-553c-4826-9dd8-c162ca665390 service nova] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Received unexpected event network-vif-plugged-d3934283-dc65-4a50-8b00-c466bb372792 for instance with vm_state building and task_state spawning. [ 1054.345045] env[62753]: DEBUG nova.compute.manager [req-75ba0c6f-bc63-4d4c-ac50-338f73660bb2 req-e8d4a152-553c-4826-9dd8-c162ca665390 service nova] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Received event network-changed-d3934283-dc65-4a50-8b00-c466bb372792 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1054.345340] env[62753]: DEBUG nova.compute.manager [req-75ba0c6f-bc63-4d4c-ac50-338f73660bb2 req-e8d4a152-553c-4826-9dd8-c162ca665390 service nova] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Refreshing instance network info cache due to event network-changed-d3934283-dc65-4a50-8b00-c466bb372792. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1054.346010] env[62753]: DEBUG oslo_concurrency.lockutils [req-75ba0c6f-bc63-4d4c-ac50-338f73660bb2 req-e8d4a152-553c-4826-9dd8-c162ca665390 service nova] Acquiring lock "refresh_cache-01d9da0b-f4e5-474c-aab2-a0f4f0c99994" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1054.413358] env[62753]: DEBUG nova.compute.manager [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1054.486617] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.686s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.490765] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.705s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.492996] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.004s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1054.497016] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.320s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1054.497016] env[62753]: DEBUG nova.objects.instance [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lazy-loading 'resources' on Instance uuid 0f81b85b-0550-46b9-b77c-9075a99cfaec {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1054.519900] env[62753]: INFO nova.scheduler.client.report [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Deleted allocations for instance ff50c5f4-ab27-4f17-948e-80e2dcc623eb [ 1054.530125] env[62753]: INFO nova.scheduler.client.report [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Deleted allocations for instance c73b1ae0-dc98-47f7-babf-e96169384785 [ 1054.608648] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1054.647053] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332741, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1054.725809] env[62753]: DEBUG nova.compute.manager [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1054.864920] env[62753]: DEBUG nova.network.neutron [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1054.938905] env[62753]: DEBUG oslo_concurrency.lockutils [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.010043] env[62753]: DEBUG nova.network.neutron [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Updating instance_info_cache with network_info: [{"id": "d3934283-dc65-4a50-8b00-c466bb372792", "address": "fa:16:3e:bf:92:e9", "network": {"id": "61c6e640-d2e6-4b3b-b191-7da9249dd339", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-703836479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8925b84dcf9a47fbaf2eb8044b3850fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3934283-dc", "ovs_interfaceid": "d3934283-dc65-4a50-8b00-c466bb372792", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1055.027023] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4558c5f9-434f-4d61-9f81-2c8261131d84 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "ff50c5f4-ab27-4f17-948e-80e2dcc623eb" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 11.344s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.041632] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bbd8eec3-2af0-49e0-8320-8cc700bb4ee2 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "c73b1ae0-dc98-47f7-babf-e96169384785" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.175s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1055.129977] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-687270ba-823c-44a9-8f01-28c5fbf587d2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.139072] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-289a5a22-3d46-466a-a571-9c4e6155c616 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.151205] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332741, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.177087] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c723afbb-37b0-4625-a6b0-ca3785cd5a5a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.184735] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cd98ddc-d6e1-4c23-a6be-24678bcfafc4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.198091] env[62753]: DEBUG nova.compute.provider_tree [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1055.254219] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1055.512598] env[62753]: DEBUG oslo_concurrency.lockutils [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Releasing lock "refresh_cache-01d9da0b-f4e5-474c-aab2-a0f4f0c99994" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1055.512980] env[62753]: DEBUG nova.compute.manager [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Instance network_info: |[{"id": "d3934283-dc65-4a50-8b00-c466bb372792", "address": "fa:16:3e:bf:92:e9", "network": {"id": "61c6e640-d2e6-4b3b-b191-7da9249dd339", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-703836479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8925b84dcf9a47fbaf2eb8044b3850fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3934283-dc", "ovs_interfaceid": "d3934283-dc65-4a50-8b00-c466bb372792", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1055.513333] env[62753]: DEBUG oslo_concurrency.lockutils [req-75ba0c6f-bc63-4d4c-ac50-338f73660bb2 req-e8d4a152-553c-4826-9dd8-c162ca665390 service nova] Acquired lock "refresh_cache-01d9da0b-f4e5-474c-aab2-a0f4f0c99994" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1055.513584] env[62753]: DEBUG nova.network.neutron [req-75ba0c6f-bc63-4d4c-ac50-338f73660bb2 req-e8d4a152-553c-4826-9dd8-c162ca665390 service nova] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Refreshing network info cache for port d3934283-dc65-4a50-8b00-c466bb372792 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1055.514775] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:92:e9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '510d3c47-3615-43d5-aa5d-a279fd915e71', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'd3934283-dc65-4a50-8b00-c466bb372792', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1055.525165] env[62753]: DEBUG oslo.service.loopingcall [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1055.526286] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1055.526884] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-c0885981-7b8b-4faa-b54a-51ebbf8a4795 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1055.548278] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1055.548278] env[62753]: value = "task-1332742" [ 1055.548278] env[62753]: _type = "Task" [ 1055.548278] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1055.557698] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332742, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.650045] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332741, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1055.701881] env[62753]: DEBUG nova.scheduler.client.report [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1055.763676] env[62753]: DEBUG nova.network.neutron [req-75ba0c6f-bc63-4d4c-ac50-338f73660bb2 req-e8d4a152-553c-4826-9dd8-c162ca665390 service nova] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Updated VIF entry in instance network info cache for port d3934283-dc65-4a50-8b00-c466bb372792. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1055.763676] env[62753]: DEBUG nova.network.neutron [req-75ba0c6f-bc63-4d4c-ac50-338f73660bb2 req-e8d4a152-553c-4826-9dd8-c162ca665390 service nova] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Updating instance_info_cache with network_info: [{"id": "d3934283-dc65-4a50-8b00-c466bb372792", "address": "fa:16:3e:bf:92:e9", "network": {"id": "61c6e640-d2e6-4b3b-b191-7da9249dd339", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-703836479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8925b84dcf9a47fbaf2eb8044b3850fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3934283-dc", "ovs_interfaceid": "d3934283-dc65-4a50-8b00-c466bb372792", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1056.061516] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332742, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.148495] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332741, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.207803] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.207803] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.209414] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.713s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.213045] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.604s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.213503] env[62753]: INFO nova.compute.claims [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1056.231753] env[62753]: INFO nova.scheduler.client.report [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Deleted allocations for instance 0f81b85b-0550-46b9-b77c-9075a99cfaec [ 1056.265729] env[62753]: DEBUG oslo_concurrency.lockutils [req-75ba0c6f-bc63-4d4c-ac50-338f73660bb2 req-e8d4a152-553c-4826-9dd8-c162ca665390 service nova] Releasing lock "refresh_cache-01d9da0b-f4e5-474c-aab2-a0f4f0c99994" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1056.559709] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332742, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.591577] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "d7ba8be1-befb-4a0b-9c35-d86e7d586203" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.591851] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "d7ba8be1-befb-4a0b-9c35-d86e7d586203" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.592080] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "d7ba8be1-befb-4a0b-9c35-d86e7d586203-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.592279] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "d7ba8be1-befb-4a0b-9c35-d86e7d586203-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.592506] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "d7ba8be1-befb-4a0b-9c35-d86e7d586203-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.594810] env[62753]: INFO nova.compute.manager [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Terminating instance [ 1056.596679] env[62753]: DEBUG nova.compute.manager [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1056.596881] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1056.597729] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18da0620-ab11-4a12-a3a4-82a6924d18bd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.605413] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1056.605707] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-770a7d53-5d84-4937-9202-ff3c845cc3f9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.613087] env[62753]: DEBUG oslo_vmware.api [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1056.613087] env[62753]: value = "task-1332743" [ 1056.613087] env[62753]: _type = "Task" [ 1056.613087] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.621278] env[62753]: DEBUG oslo_vmware.api [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332743, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.649311] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332741, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.664075] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "961ced74-10af-4443-857e-545cd890674e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.664356] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "961ced74-10af-4443-857e-545cd890674e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.664628] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "961ced74-10af-4443-857e-545cd890674e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1056.664827] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "961ced74-10af-4443-857e-545cd890674e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1056.665012] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "961ced74-10af-4443-857e-545cd890674e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1056.667387] env[62753]: INFO nova.compute.manager [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Terminating instance [ 1056.669478] env[62753]: DEBUG nova.compute.manager [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1056.669684] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1056.670558] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f11e47c3-4d07-4fd2-a62f-ba92dc78bb7b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.678826] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1056.679152] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-761a2f83-54ea-411a-9742-20261b41403f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1056.685594] env[62753]: DEBUG oslo_vmware.api [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1056.685594] env[62753]: value = "task-1332744" [ 1056.685594] env[62753]: _type = "Task" [ 1056.685594] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1056.694405] env[62753]: DEBUG oslo_vmware.api [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332744, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1056.710244] env[62753]: DEBUG nova.compute.manager [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1056.742049] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cb95e8e4-2fae-41c2-a0f4-eeb7b1027d75 tempest-ServerDiskConfigTestJSON-1149266675 tempest-ServerDiskConfigTestJSON-1149266675-project-member] Lock "0f81b85b-0550-46b9-b77c-9075a99cfaec" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.520s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1057.061152] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332742, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.122916] env[62753]: DEBUG oslo_vmware.api [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332743, 'name': PowerOffVM_Task, 'duration_secs': 0.228562} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.123242] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1057.123420] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1057.123724] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-6f03a150-304e-4d1a-a1c5-b7bbd3cf285e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.151903] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332741, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.191270] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1057.191497] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1057.191689] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Deleting the datastore file [datastore2] d7ba8be1-befb-4a0b-9c35-d86e7d586203 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1057.192389] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-99798b07-f56d-456d-9f40-44e6672cff8c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.198306] env[62753]: DEBUG oslo_vmware.api [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332744, 'name': PowerOffVM_Task, 'duration_secs': 0.185745} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.199195] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1057.199465] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1057.199777] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-4a90ba3a-866a-4b8a-bd1d-3cca1ec27b43 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.202675] env[62753]: DEBUG oslo_vmware.api [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1057.202675] env[62753]: value = "task-1332746" [ 1057.202675] env[62753]: _type = "Task" [ 1057.202675] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.211357] env[62753]: DEBUG oslo_vmware.api [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332746, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.239742] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1057.264063] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1057.264343] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1057.264615] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Deleting the datastore file [datastore2] 961ced74-10af-4443-857e-545cd890674e {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1057.265088] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-1a490c10-ee25-4063-96df-b1c30121d5d4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.270818] env[62753]: DEBUG oslo_vmware.api [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for the task: (returnval){ [ 1057.270818] env[62753]: value = "task-1332748" [ 1057.270818] env[62753]: _type = "Task" [ 1057.270818] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1057.282342] env[62753]: DEBUG oslo_vmware.api [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332748, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.370238] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47211970-c787-4c8d-9539-2105135af8c7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.378403] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-60bd6d29-9b0c-493f-9d04-02bd369aedc6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.410928] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3814fde9-075d-4b6a-aa75-17a16e5f84ef {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.417853] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d2c24d7e-4bde-41c9-aa12-437025b9c3d3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.433981] env[62753]: DEBUG nova.compute.provider_tree [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1057.561524] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332742, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1057.650471] env[62753]: DEBUG oslo_vmware.api [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332741, 'name': PowerOnVM_Task, 'duration_secs': 4.0159} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.651169] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1057.651169] env[62753]: INFO nova.compute.manager [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Took 11.64 seconds to spawn the instance on the hypervisor. [ 1057.651449] env[62753]: DEBUG nova.compute.manager [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1057.652554] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3876d5b5-6a1b-46fb-a0cb-7c32538e5bde {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1057.714974] env[62753]: DEBUG oslo_vmware.api [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332746, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.185319} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.715299] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1057.715504] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1057.715687] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1057.715866] env[62753]: INFO nova.compute.manager [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1057.716138] env[62753]: DEBUG oslo.service.loopingcall [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1057.716338] env[62753]: DEBUG nova.compute.manager [-] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1057.716434] env[62753]: DEBUG nova.network.neutron [-] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1057.783836] env[62753]: DEBUG oslo_vmware.api [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Task: {'id': task-1332748, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.192774} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1057.784382] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1057.784382] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1057.784601] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1057.784838] env[62753]: INFO nova.compute.manager [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] [instance: 961ced74-10af-4443-857e-545cd890674e] Took 1.12 seconds to destroy the instance on the hypervisor. [ 1057.786084] env[62753]: DEBUG oslo.service.loopingcall [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1057.786084] env[62753]: DEBUG nova.compute.manager [-] [instance: 961ced74-10af-4443-857e-545cd890674e] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1057.786084] env[62753]: DEBUG nova.network.neutron [-] [instance: 961ced74-10af-4443-857e-545cd890674e] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1057.940632] env[62753]: DEBUG nova.scheduler.client.report [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1058.065414] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332742, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.156093] env[62753]: DEBUG nova.compute.manager [req-f81e4837-76b2-48e6-b32d-87bd68e41a26 req-3591cb0a-fcbb-436e-b8ee-91815a3355f8 service nova] [instance: 961ced74-10af-4443-857e-545cd890674e] Received event network-vif-deleted-6e141f4e-b7cc-491d-b299-8ae721ee06e5 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1058.156093] env[62753]: INFO nova.compute.manager [req-f81e4837-76b2-48e6-b32d-87bd68e41a26 req-3591cb0a-fcbb-436e-b8ee-91815a3355f8 service nova] [instance: 961ced74-10af-4443-857e-545cd890674e] Neutron deleted interface 6e141f4e-b7cc-491d-b299-8ae721ee06e5; detaching it from the instance and deleting it from the info cache [ 1058.156093] env[62753]: DEBUG nova.network.neutron [req-f81e4837-76b2-48e6-b32d-87bd68e41a26 req-3591cb0a-fcbb-436e-b8ee-91815a3355f8 service nova] [instance: 961ced74-10af-4443-857e-545cd890674e] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.174707] env[62753]: INFO nova.compute.manager [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Took 22.09 seconds to build instance. [ 1058.448200] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.236s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.448731] env[62753]: DEBUG nova.compute.manager [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1058.452491] env[62753]: DEBUG oslo_concurrency.lockutils [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 3.513s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1058.453480] env[62753]: INFO nova.compute.claims [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1058.564200] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332742, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1058.619346] env[62753]: DEBUG nova.network.neutron [-] [instance: 961ced74-10af-4443-857e-545cd890674e] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.638451] env[62753]: DEBUG nova.network.neutron [-] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1058.658710] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-0719be67-e410-448c-a825-2788c52aa877 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.670149] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f5b89717-aea8-4d11-9b2f-a81b9000aebe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1058.682949] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fe989f06-4f6e-4c71-946e-6f4c861fadd4 tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.608s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1058.699653] env[62753]: DEBUG nova.compute.manager [req-f81e4837-76b2-48e6-b32d-87bd68e41a26 req-3591cb0a-fcbb-436e-b8ee-91815a3355f8 service nova] [instance: 961ced74-10af-4443-857e-545cd890674e] Detach interface failed, port_id=6e141f4e-b7cc-491d-b299-8ae721ee06e5, reason: Instance 961ced74-10af-4443-857e-545cd890674e could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1058.961932] env[62753]: DEBUG nova.compute.utils [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1058.963471] env[62753]: DEBUG nova.compute.manager [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1058.963648] env[62753]: DEBUG nova.network.neutron [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1059.016075] env[62753]: DEBUG nova.policy [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '128e8db9dcaf437a9dc1a7541d335223', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b76b178732e41688f767fdd65934193', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1059.062774] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332742, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.120965] env[62753]: INFO nova.compute.manager [-] [instance: 961ced74-10af-4443-857e-545cd890674e] Took 1.34 seconds to deallocate network for instance. [ 1059.142023] env[62753]: INFO nova.compute.manager [-] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Took 1.42 seconds to deallocate network for instance. [ 1059.267484] env[62753]: DEBUG nova.network.neutron [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Successfully created port: 2254df7c-8da6-4c8d-ae9c-5b10e79dc88f {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1059.416191] env[62753]: DEBUG nova.compute.manager [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Stashing vm_state: active {{(pid=62753) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1059.471559] env[62753]: DEBUG nova.compute.manager [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1059.564860] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332742, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1059.614785] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f70b5065-348a-4c65-9c28-10a6ed94eee4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.622467] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4315d754-5436-4f57-9119-0881d1113f46 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.628087] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.656047] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1059.656877] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-68cf8e7a-5d88-4d8c-aad6-b9064a562c1b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.664347] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3444e83-5b78-4035-81cd-706419fd1101 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1059.677655] env[62753]: DEBUG nova.compute.provider_tree [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1059.939758] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1060.065577] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332742, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.180834] env[62753]: DEBUG nova.scheduler.client.report [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1060.188175] env[62753]: DEBUG nova.compute.manager [req-53b31b10-b88d-4657-8084-794a75f80006 req-99f7693e-eb6b-4def-9664-d86fe979657d service nova] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Received event network-vif-deleted-0c86bac1-aaf9-4933-ac5a-6488554c390c {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1060.485903] env[62753]: DEBUG nova.compute.manager [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1060.517349] env[62753]: DEBUG nova.virt.hardware [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1060.517606] env[62753]: DEBUG nova.virt.hardware [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1060.517771] env[62753]: DEBUG nova.virt.hardware [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1060.517957] env[62753]: DEBUG nova.virt.hardware [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1060.518124] env[62753]: DEBUG nova.virt.hardware [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1060.518280] env[62753]: DEBUG nova.virt.hardware [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1060.518495] env[62753]: DEBUG nova.virt.hardware [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1060.518659] env[62753]: DEBUG nova.virt.hardware [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1060.518835] env[62753]: DEBUG nova.virt.hardware [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1060.519008] env[62753]: DEBUG nova.virt.hardware [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1060.519199] env[62753]: DEBUG nova.virt.hardware [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1060.520068] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e42684af-36b1-4068-909f-82b078536c31 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.527981] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7b08e314-e97c-4ecb-8c18-2d3b58a71cc6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.562221] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332742, 'name': CreateVM_Task, 'duration_secs': 4.833941} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1060.562401] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1060.563092] env[62753]: DEBUG oslo_concurrency.lockutils [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1060.563264] env[62753]: DEBUG oslo_concurrency.lockutils [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1060.563602] env[62753]: DEBUG oslo_concurrency.lockutils [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1060.563824] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-81f65bb5-bc73-4537-a23a-5c67de11ce34 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1060.568280] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1060.568280] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]528726ab-8bd6-9806-3c8d-74a1ad262719" [ 1060.568280] env[62753]: _type = "Task" [ 1060.568280] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1060.576223] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]528726ab-8bd6-9806-3c8d-74a1ad262719, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1060.687580] env[62753]: DEBUG oslo_concurrency.lockutils [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.235s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1060.688155] env[62753]: DEBUG nova.compute.manager [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1060.691718] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.438s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1060.694439] env[62753]: INFO nova.compute.claims [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1060.957976] env[62753]: DEBUG nova.network.neutron [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Successfully updated port: 2254df7c-8da6-4c8d-ae9c-5b10e79dc88f {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1061.079063] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]528726ab-8bd6-9806-3c8d-74a1ad262719, 'name': SearchDatastore_Task, 'duration_secs': 0.040889} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.079459] env[62753]: DEBUG oslo_concurrency.lockutils [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1061.079643] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1061.079883] env[62753]: DEBUG oslo_concurrency.lockutils [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.080043] env[62753]: DEBUG oslo_concurrency.lockutils [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.080331] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1061.080514] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-25b9ab2f-5dcf-42cf-b920-032fa5ba42a7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.091081] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1061.091351] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1061.092123] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-666c1dc3-12c7-4ec8-bf00-998572d9a8d1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.099302] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1061.099302] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f9f57a-fec9-390e-e950-5f070047ce0b" [ 1061.099302] env[62753]: _type = "Task" [ 1061.099302] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.106580] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f9f57a-fec9-390e-e950-5f070047ce0b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.200509] env[62753]: DEBUG nova.compute.utils [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1061.203696] env[62753]: DEBUG nova.compute.manager [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1061.203863] env[62753]: DEBUG nova.network.neutron [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1061.272594] env[62753]: DEBUG nova.policy [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '128e8db9dcaf437a9dc1a7541d335223', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b76b178732e41688f767fdd65934193', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1061.460604] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "refresh_cache-6e417227-895d-4576-b025-7d4a0aafa379" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1061.460604] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquired lock "refresh_cache-6e417227-895d-4576-b025-7d4a0aafa379" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1061.460604] env[62753]: DEBUG nova.network.neutron [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1061.610989] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f9f57a-fec9-390e-e950-5f070047ce0b, 'name': SearchDatastore_Task, 'duration_secs': 0.030025} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1061.611841] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ff11aab1-d1c4-4aff-a49b-96de83da150a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.618099] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1061.618099] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ff64f8-0bd6-a32f-f1aa-38365da58565" [ 1061.618099] env[62753]: _type = "Task" [ 1061.618099] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1061.626063] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ff64f8-0bd6-a32f-f1aa-38365da58565, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1061.690430] env[62753]: DEBUG nova.network.neutron [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Successfully created port: 637f03b2-1077-4b59-98b1-a33ff9e7402b {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1061.705134] env[62753]: DEBUG nova.compute.manager [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1061.863212] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-919f4cb0-a1b2-4c79-81eb-1de4e9b8df41 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.871010] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe7c0917-5c6c-407d-b6eb-2a2757aecb0d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.905654] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b5c63b40-0b53-4437-bd68-4507dee75527 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.913161] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f6fb6133-0570-4042-bfe6-6d51894c5949 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1061.927801] env[62753]: DEBUG nova.compute.provider_tree [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1061.992521] env[62753]: DEBUG nova.network.neutron [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1062.129155] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ff64f8-0bd6-a32f-f1aa-38365da58565, 'name': SearchDatastore_Task, 'duration_secs': 0.009824} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1062.133021] env[62753]: DEBUG oslo_concurrency.lockutils [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.133021] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 01d9da0b-f4e5-474c-aab2-a0f4f0c99994/01d9da0b-f4e5-474c-aab2-a0f4f0c99994.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1062.133021] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-56c692a4-ea92-4e4a-8b99-ef88a30de454 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.136433] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1062.136433] env[62753]: value = "task-1332749" [ 1062.136433] env[62753]: _type = "Task" [ 1062.136433] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.144912] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332749, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.146299] env[62753]: DEBUG nova.network.neutron [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Updating instance_info_cache with network_info: [{"id": "2254df7c-8da6-4c8d-ae9c-5b10e79dc88f", "address": "fa:16:3e:05:70:9f", "network": {"id": "e3de96b6-0e52-4885-8747-d3e60da74ef2", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-91543986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b76b178732e41688f767fdd65934193", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2254df7c-8d", "ovs_interfaceid": "2254df7c-8da6-4c8d-ae9c-5b10e79dc88f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1062.210685] env[62753]: DEBUG nova.compute.manager [req-31ca8b06-4365-4e17-96d2-8392368c9615 req-2603e628-0185-4a34-9369-19fabf55120a service nova] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Received event network-vif-plugged-2254df7c-8da6-4c8d-ae9c-5b10e79dc88f {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1062.210960] env[62753]: DEBUG oslo_concurrency.lockutils [req-31ca8b06-4365-4e17-96d2-8392368c9615 req-2603e628-0185-4a34-9369-19fabf55120a service nova] Acquiring lock "6e417227-895d-4576-b025-7d4a0aafa379-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1062.211412] env[62753]: DEBUG oslo_concurrency.lockutils [req-31ca8b06-4365-4e17-96d2-8392368c9615 req-2603e628-0185-4a34-9369-19fabf55120a service nova] Lock "6e417227-895d-4576-b025-7d4a0aafa379-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.211643] env[62753]: DEBUG oslo_concurrency.lockutils [req-31ca8b06-4365-4e17-96d2-8392368c9615 req-2603e628-0185-4a34-9369-19fabf55120a service nova] Lock "6e417227-895d-4576-b025-7d4a0aafa379-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.211884] env[62753]: DEBUG nova.compute.manager [req-31ca8b06-4365-4e17-96d2-8392368c9615 req-2603e628-0185-4a34-9369-19fabf55120a service nova] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] No waiting events found dispatching network-vif-plugged-2254df7c-8da6-4c8d-ae9c-5b10e79dc88f {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1062.212175] env[62753]: WARNING nova.compute.manager [req-31ca8b06-4365-4e17-96d2-8392368c9615 req-2603e628-0185-4a34-9369-19fabf55120a service nova] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Received unexpected event network-vif-plugged-2254df7c-8da6-4c8d-ae9c-5b10e79dc88f for instance with vm_state building and task_state spawning. [ 1062.212412] env[62753]: DEBUG nova.compute.manager [req-31ca8b06-4365-4e17-96d2-8392368c9615 req-2603e628-0185-4a34-9369-19fabf55120a service nova] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Received event network-changed-2254df7c-8da6-4c8d-ae9c-5b10e79dc88f {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1062.212769] env[62753]: DEBUG nova.compute.manager [req-31ca8b06-4365-4e17-96d2-8392368c9615 req-2603e628-0185-4a34-9369-19fabf55120a service nova] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Refreshing instance network info cache due to event network-changed-2254df7c-8da6-4c8d-ae9c-5b10e79dc88f. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1062.213105] env[62753]: DEBUG oslo_concurrency.lockutils [req-31ca8b06-4365-4e17-96d2-8392368c9615 req-2603e628-0185-4a34-9369-19fabf55120a service nova] Acquiring lock "refresh_cache-6e417227-895d-4576-b025-7d4a0aafa379" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1062.430892] env[62753]: DEBUG nova.scheduler.client.report [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1062.648202] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332749, 'name': CopyVirtualDisk_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.648663] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Releasing lock "refresh_cache-6e417227-895d-4576-b025-7d4a0aafa379" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1062.648961] env[62753]: DEBUG nova.compute.manager [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Instance network_info: |[{"id": "2254df7c-8da6-4c8d-ae9c-5b10e79dc88f", "address": "fa:16:3e:05:70:9f", "network": {"id": "e3de96b6-0e52-4885-8747-d3e60da74ef2", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-91543986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b76b178732e41688f767fdd65934193", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2254df7c-8d", "ovs_interfaceid": "2254df7c-8da6-4c8d-ae9c-5b10e79dc88f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1062.649259] env[62753]: DEBUG oslo_concurrency.lockutils [req-31ca8b06-4365-4e17-96d2-8392368c9615 req-2603e628-0185-4a34-9369-19fabf55120a service nova] Acquired lock "refresh_cache-6e417227-895d-4576-b025-7d4a0aafa379" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1062.649438] env[62753]: DEBUG nova.network.neutron [req-31ca8b06-4365-4e17-96d2-8392368c9615 req-2603e628-0185-4a34-9369-19fabf55120a service nova] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Refreshing network info cache for port 2254df7c-8da6-4c8d-ae9c-5b10e79dc88f {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1062.650778] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:05:70:9f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '13d625c9-77ec-4edb-a56b-9f37a314cc39', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '2254df7c-8da6-4c8d-ae9c-5b10e79dc88f', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1062.658547] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Creating folder: Project (7b76b178732e41688f767fdd65934193). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1062.659062] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-63ac87d1-3999-481d-9204-87520bb65ec3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.669608] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Created folder: Project (7b76b178732e41688f767fdd65934193) in parent group-v284541. [ 1062.669828] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Creating folder: Instances. Parent ref: group-v284676. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1062.670105] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-35c8b9eb-b25c-4d96-8b7b-201c963d9fea {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.678813] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Created folder: Instances in parent group-v284676. [ 1062.679092] env[62753]: DEBUG oslo.service.loopingcall [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1062.679312] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1062.679565] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-3b37ec3f-28d3-43a0-97bf-6bb35e8cbf57 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.699856] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1062.699856] env[62753]: value = "task-1332752" [ 1062.699856] env[62753]: _type = "Task" [ 1062.699856] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1062.711496] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332752, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1062.718949] env[62753]: DEBUG nova.compute.manager [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1062.746701] env[62753]: DEBUG nova.virt.hardware [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1062.747012] env[62753]: DEBUG nova.virt.hardware [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1062.747445] env[62753]: DEBUG nova.virt.hardware [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1062.747714] env[62753]: DEBUG nova.virt.hardware [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1062.747930] env[62753]: DEBUG nova.virt.hardware [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1062.749405] env[62753]: DEBUG nova.virt.hardware [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1062.749405] env[62753]: DEBUG nova.virt.hardware [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1062.749405] env[62753]: DEBUG nova.virt.hardware [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1062.749405] env[62753]: DEBUG nova.virt.hardware [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1062.749405] env[62753]: DEBUG nova.virt.hardware [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1062.749405] env[62753]: DEBUG nova.virt.hardware [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1062.750639] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e7f49d8-f28e-429d-bacc-2d3fed3181d5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.758674] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21a87f9d-3701-4403-8a5f-98a85f33ff21 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1062.936635] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.245s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1062.937275] env[62753]: DEBUG nova.compute.manager [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1062.939964] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 5.700s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1062.941583] env[62753]: INFO nova.compute.claims [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1063.152270] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332749, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.544232} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.153235] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 01d9da0b-f4e5-474c-aab2-a0f4f0c99994/01d9da0b-f4e5-474c-aab2-a0f4f0c99994.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1063.153235] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1063.153235] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-dab0f96e-b044-4532-87f1-e7b9ea86c056 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.161677] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1063.161677] env[62753]: value = "task-1332753" [ 1063.161677] env[62753]: _type = "Task" [ 1063.161677] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.169700] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332753, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.210683] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332752, 'name': CreateVM_Task, 'duration_secs': 0.297202} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.210857] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1063.211539] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.211712] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.212052] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1063.212360] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-14f00803-1667-4950-befc-eb57afe768f9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.217181] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1063.217181] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c41cc9-d381-b175-ba92-5278a4003e93" [ 1063.217181] env[62753]: _type = "Task" [ 1063.217181] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.224733] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c41cc9-d381-b175-ba92-5278a4003e93, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.264048] env[62753]: DEBUG nova.network.neutron [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Successfully updated port: 637f03b2-1077-4b59-98b1-a33ff9e7402b {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1063.391501] env[62753]: DEBUG nova.network.neutron [req-31ca8b06-4365-4e17-96d2-8392368c9615 req-2603e628-0185-4a34-9369-19fabf55120a service nova] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Updated VIF entry in instance network info cache for port 2254df7c-8da6-4c8d-ae9c-5b10e79dc88f. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1063.391859] env[62753]: DEBUG nova.network.neutron [req-31ca8b06-4365-4e17-96d2-8392368c9615 req-2603e628-0185-4a34-9369-19fabf55120a service nova] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Updating instance_info_cache with network_info: [{"id": "2254df7c-8da6-4c8d-ae9c-5b10e79dc88f", "address": "fa:16:3e:05:70:9f", "network": {"id": "e3de96b6-0e52-4885-8747-d3e60da74ef2", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-91543986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b76b178732e41688f767fdd65934193", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2254df7c-8d", "ovs_interfaceid": "2254df7c-8da6-4c8d-ae9c-5b10e79dc88f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1063.448264] env[62753]: DEBUG nova.compute.utils [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1063.449601] env[62753]: DEBUG nova.compute.manager [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1063.449825] env[62753]: DEBUG nova.network.neutron [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1063.494023] env[62753]: DEBUG nova.policy [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '128e8db9dcaf437a9dc1a7541d335223', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '7b76b178732e41688f767fdd65934193', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1063.644831] env[62753]: DEBUG oslo_concurrency.lockutils [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "22cea165-cfde-403f-a0e2-82a86f5baa51" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1063.645083] env[62753]: DEBUG oslo_concurrency.lockutils [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1063.672612] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332753, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.734299] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c41cc9-d381-b175-ba92-5278a4003e93, 'name': SearchDatastore_Task, 'duration_secs': 0.052359} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1063.734299] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.734299] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1063.734299] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.734299] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.734299] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1063.734299] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0d03b09c-f3bd-4a33-8b02-e84137e1acbc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.742101] env[62753]: DEBUG nova.network.neutron [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Successfully created port: 3f1e92c2-bb9a-4d5d-a2ea-8f5943d15260 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1063.745018] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1063.745411] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1063.747018] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bd2befd0-7d3a-40a3-a5f8-4d290c07ec9e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1063.755024] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1063.755024] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]524b9340-e0c4-4367-4bbc-3e951d3fc6d6" [ 1063.755024] env[62753]: _type = "Task" [ 1063.755024] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1063.760676] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]524b9340-e0c4-4367-4bbc-3e951d3fc6d6, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1063.767623] env[62753]: DEBUG oslo_concurrency.lockutils [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "refresh_cache-32563e1f-398f-4fc9-866f-d85d9f3f27a7" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1063.768087] env[62753]: DEBUG oslo_concurrency.lockutils [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquired lock "refresh_cache-32563e1f-398f-4fc9-866f-d85d9f3f27a7" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1063.768324] env[62753]: DEBUG nova.network.neutron [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1063.894842] env[62753]: DEBUG oslo_concurrency.lockutils [req-31ca8b06-4365-4e17-96d2-8392368c9615 req-2603e628-0185-4a34-9369-19fabf55120a service nova] Releasing lock "refresh_cache-6e417227-895d-4576-b025-7d4a0aafa379" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1063.955592] env[62753]: DEBUG nova.compute.manager [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1064.113177] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a36256a9-00fc-482c-baf0-ac804f05cc8f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.120922] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8016b0fa-20c7-45a5-a86f-d7920dfe06da {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.152429] env[62753]: DEBUG nova.compute.manager [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1064.155851] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d06ef173-7658-40ad-9184-fd9f4c3edec4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.167238] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27d4e4ef-7630-4043-9276-ae3430c27330 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.177949] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332753, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.64704} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.185577] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1064.186120] env[62753]: DEBUG nova.compute.provider_tree [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1064.188482] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c60e2f2-393b-4cbb-a12e-ecbe1cc0cde2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.211567] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Reconfiguring VM instance instance-00000063 to attach disk [datastore1] 01d9da0b-f4e5-474c-aab2-a0f4f0c99994/01d9da0b-f4e5-474c-aab2-a0f4f0c99994.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1064.212087] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6dc16600-e16a-4931-b952-4efd61a47e6e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.231409] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1064.231409] env[62753]: value = "task-1332754" [ 1064.231409] env[62753]: _type = "Task" [ 1064.231409] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.236214] env[62753]: DEBUG nova.compute.manager [req-181f0a40-7147-4931-a64b-f3a21250edb0 req-cdbe0f54-25a8-46ee-b6fa-7252995194e1 service nova] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Received event network-vif-plugged-637f03b2-1077-4b59-98b1-a33ff9e7402b {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1064.236420] env[62753]: DEBUG oslo_concurrency.lockutils [req-181f0a40-7147-4931-a64b-f3a21250edb0 req-cdbe0f54-25a8-46ee-b6fa-7252995194e1 service nova] Acquiring lock "32563e1f-398f-4fc9-866f-d85d9f3f27a7-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.236629] env[62753]: DEBUG oslo_concurrency.lockutils [req-181f0a40-7147-4931-a64b-f3a21250edb0 req-cdbe0f54-25a8-46ee-b6fa-7252995194e1 service nova] Lock "32563e1f-398f-4fc9-866f-d85d9f3f27a7-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1064.236797] env[62753]: DEBUG oslo_concurrency.lockutils [req-181f0a40-7147-4931-a64b-f3a21250edb0 req-cdbe0f54-25a8-46ee-b6fa-7252995194e1 service nova] Lock "32563e1f-398f-4fc9-866f-d85d9f3f27a7-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1064.236970] env[62753]: DEBUG nova.compute.manager [req-181f0a40-7147-4931-a64b-f3a21250edb0 req-cdbe0f54-25a8-46ee-b6fa-7252995194e1 service nova] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] No waiting events found dispatching network-vif-plugged-637f03b2-1077-4b59-98b1-a33ff9e7402b {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1064.237221] env[62753]: WARNING nova.compute.manager [req-181f0a40-7147-4931-a64b-f3a21250edb0 req-cdbe0f54-25a8-46ee-b6fa-7252995194e1 service nova] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Received unexpected event network-vif-plugged-637f03b2-1077-4b59-98b1-a33ff9e7402b for instance with vm_state building and task_state spawning. [ 1064.237410] env[62753]: DEBUG nova.compute.manager [req-181f0a40-7147-4931-a64b-f3a21250edb0 req-cdbe0f54-25a8-46ee-b6fa-7252995194e1 service nova] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Received event network-changed-637f03b2-1077-4b59-98b1-a33ff9e7402b {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1064.237571] env[62753]: DEBUG nova.compute.manager [req-181f0a40-7147-4931-a64b-f3a21250edb0 req-cdbe0f54-25a8-46ee-b6fa-7252995194e1 service nova] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Refreshing instance network info cache due to event network-changed-637f03b2-1077-4b59-98b1-a33ff9e7402b. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1064.237775] env[62753]: DEBUG oslo_concurrency.lockutils [req-181f0a40-7147-4931-a64b-f3a21250edb0 req-cdbe0f54-25a8-46ee-b6fa-7252995194e1 service nova] Acquiring lock "refresh_cache-32563e1f-398f-4fc9-866f-d85d9f3f27a7" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1064.243968] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332754, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.262517] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]524b9340-e0c4-4367-4bbc-3e951d3fc6d6, 'name': SearchDatastore_Task, 'duration_secs': 0.00969} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.262997] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-696a295a-44e4-40bc-8163-bdac28f4a0b8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.267661] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1064.267661] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a0fce4-aef5-c374-8a71-5b40faefb98a" [ 1064.267661] env[62753]: _type = "Task" [ 1064.267661] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.276780] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a0fce4-aef5-c374-8a71-5b40faefb98a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.305027] env[62753]: DEBUG nova.network.neutron [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1064.436616] env[62753]: DEBUG nova.network.neutron [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Updating instance_info_cache with network_info: [{"id": "637f03b2-1077-4b59-98b1-a33ff9e7402b", "address": "fa:16:3e:0e:f4:3f", "network": {"id": "e3de96b6-0e52-4885-8747-d3e60da74ef2", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-91543986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b76b178732e41688f767fdd65934193", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap637f03b2-10", "ovs_interfaceid": "637f03b2-1077-4b59-98b1-a33ff9e7402b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1064.673847] env[62753]: DEBUG oslo_concurrency.lockutils [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1064.691975] env[62753]: DEBUG nova.scheduler.client.report [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1064.741949] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332754, 'name': ReconfigVM_Task, 'duration_secs': 0.267941} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.742249] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Reconfigured VM instance instance-00000063 to attach disk [datastore1] 01d9da0b-f4e5-474c-aab2-a0f4f0c99994/01d9da0b-f4e5-474c-aab2-a0f4f0c99994.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1064.742943] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ed1993b0-c014-4e7d-87ad-eaef3ac3e148 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.748597] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1064.748597] env[62753]: value = "task-1332755" [ 1064.748597] env[62753]: _type = "Task" [ 1064.748597] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.756968] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332755, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.777118] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a0fce4-aef5-c374-8a71-5b40faefb98a, 'name': SearchDatastore_Task, 'duration_secs': 0.009103} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1064.777451] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.777769] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 6e417227-895d-4576-b025-7d4a0aafa379/6e417227-895d-4576-b025-7d4a0aafa379.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1064.778091] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a73f3033-5cf0-44a2-b85d-0a614e3f7d02 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.784713] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1064.784713] env[62753]: value = "task-1332756" [ 1064.784713] env[62753]: _type = "Task" [ 1064.784713] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.792119] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332756, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1064.940112] env[62753]: DEBUG oslo_concurrency.lockutils [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Releasing lock "refresh_cache-32563e1f-398f-4fc9-866f-d85d9f3f27a7" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1064.940440] env[62753]: DEBUG nova.compute.manager [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Instance network_info: |[{"id": "637f03b2-1077-4b59-98b1-a33ff9e7402b", "address": "fa:16:3e:0e:f4:3f", "network": {"id": "e3de96b6-0e52-4885-8747-d3e60da74ef2", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-91543986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b76b178732e41688f767fdd65934193", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap637f03b2-10", "ovs_interfaceid": "637f03b2-1077-4b59-98b1-a33ff9e7402b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1064.940808] env[62753]: DEBUG oslo_concurrency.lockutils [req-181f0a40-7147-4931-a64b-f3a21250edb0 req-cdbe0f54-25a8-46ee-b6fa-7252995194e1 service nova] Acquired lock "refresh_cache-32563e1f-398f-4fc9-866f-d85d9f3f27a7" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1064.941023] env[62753]: DEBUG nova.network.neutron [req-181f0a40-7147-4931-a64b-f3a21250edb0 req-cdbe0f54-25a8-46ee-b6fa-7252995194e1 service nova] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Refreshing network info cache for port 637f03b2-1077-4b59-98b1-a33ff9e7402b {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1064.942355] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:0e:f4:3f', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '13d625c9-77ec-4edb-a56b-9f37a314cc39', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '637f03b2-1077-4b59-98b1-a33ff9e7402b', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1064.949783] env[62753]: DEBUG oslo.service.loopingcall [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1064.952696] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1064.955147] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-01edc7d5-8a6a-41c0-b13c-f1bb46f99400 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1064.970624] env[62753]: DEBUG nova.compute.manager [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1064.980500] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1064.980500] env[62753]: value = "task-1332757" [ 1064.980500] env[62753]: _type = "Task" [ 1064.980500] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1064.991603] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332757, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.001383] env[62753]: DEBUG nova.virt.hardware [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1065.001667] env[62753]: DEBUG nova.virt.hardware [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1065.001889] env[62753]: DEBUG nova.virt.hardware [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1065.002139] env[62753]: DEBUG nova.virt.hardware [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1065.002300] env[62753]: DEBUG nova.virt.hardware [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1065.002456] env[62753]: DEBUG nova.virt.hardware [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1065.002680] env[62753]: DEBUG nova.virt.hardware [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1065.002847] env[62753]: DEBUG nova.virt.hardware [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1065.003035] env[62753]: DEBUG nova.virt.hardware [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1065.003216] env[62753]: DEBUG nova.virt.hardware [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1065.003397] env[62753]: DEBUG nova.virt.hardware [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1065.004906] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1a0a9e90-3dcc-4ea4-ad16-f4f1a9f2ea56 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.014234] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d017dea0-f74d-4e6c-bec1-f7304442c3cf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.196751] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.257s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1065.197321] env[62753]: DEBUG nova.compute.manager [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1065.200471] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 5.572s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1065.200471] env[62753]: DEBUG nova.objects.instance [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lazy-loading 'resources' on Instance uuid 961ced74-10af-4443-857e-545cd890674e {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1065.261166] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332755, 'name': Rename_Task, 'duration_secs': 0.185432} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.264179] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1065.264552] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-acfb12ce-fedb-40cd-a150-4efa4e3e87f6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.272131] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1065.272131] env[62753]: value = "task-1332758" [ 1065.272131] env[62753]: _type = "Task" [ 1065.272131] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.280245] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332758, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.294592] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332756, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.498572} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.294880] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 6e417227-895d-4576-b025-7d4a0aafa379/6e417227-895d-4576-b025-7d4a0aafa379.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1065.295712] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1065.295712] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-e746f76f-c829-4d94-82b0-ebadd69717b1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.302052] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1065.302052] env[62753]: value = "task-1332759" [ 1065.302052] env[62753]: _type = "Task" [ 1065.302052] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.310498] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332759, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.327571] env[62753]: DEBUG nova.network.neutron [req-181f0a40-7147-4931-a64b-f3a21250edb0 req-cdbe0f54-25a8-46ee-b6fa-7252995194e1 service nova] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Updated VIF entry in instance network info cache for port 637f03b2-1077-4b59-98b1-a33ff9e7402b. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1065.327989] env[62753]: DEBUG nova.network.neutron [req-181f0a40-7147-4931-a64b-f3a21250edb0 req-cdbe0f54-25a8-46ee-b6fa-7252995194e1 service nova] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Updating instance_info_cache with network_info: [{"id": "637f03b2-1077-4b59-98b1-a33ff9e7402b", "address": "fa:16:3e:0e:f4:3f", "network": {"id": "e3de96b6-0e52-4885-8747-d3e60da74ef2", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-91543986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b76b178732e41688f767fdd65934193", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap637f03b2-10", "ovs_interfaceid": "637f03b2-1077-4b59-98b1-a33ff9e7402b", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1065.430228] env[62753]: DEBUG nova.network.neutron [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Successfully updated port: 3f1e92c2-bb9a-4d5d-a2ea-8f5943d15260 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1065.490130] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332757, 'name': CreateVM_Task, 'duration_secs': 0.498516} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.490303] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1065.490990] env[62753]: DEBUG oslo_concurrency.lockutils [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.491184] env[62753]: DEBUG oslo_concurrency.lockutils [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.491515] env[62753]: DEBUG oslo_concurrency.lockutils [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1065.491776] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8984a2b9-172f-4e91-8e2f-11f12897afcb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.496361] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1065.496361] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52103667-677c-a9c9-5415-e02793a87abf" [ 1065.496361] env[62753]: _type = "Task" [ 1065.496361] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.503763] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52103667-677c-a9c9-5415-e02793a87abf, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.703992] env[62753]: DEBUG nova.compute.utils [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1065.709278] env[62753]: DEBUG nova.compute.manager [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1065.709412] env[62753]: DEBUG nova.network.neutron [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1065.751185] env[62753]: DEBUG nova.policy [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'af45e909b0c84662a8cfdf92896e5fe5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '196457eaa51643c38b4ef4fc6fbbcba3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1065.785206] env[62753]: DEBUG oslo_vmware.api [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332758, 'name': PowerOnVM_Task, 'duration_secs': 0.469955} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.787898] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1065.787898] env[62753]: INFO nova.compute.manager [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Took 12.73 seconds to spawn the instance on the hypervisor. [ 1065.787898] env[62753]: DEBUG nova.compute.manager [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1065.790768] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-427a42d3-002d-40a1-bb82-d7551975c5fd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.814465] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332759, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.125405} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1065.814465] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1065.815308] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ac048cb7-ee94-4f72-8dd8-c7fa01ab1f7f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.837213] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Reconfiguring VM instance instance-00000064 to attach disk [datastore2] 6e417227-895d-4576-b025-7d4a0aafa379/6e417227-895d-4576-b025-7d4a0aafa379.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1065.841109] env[62753]: DEBUG oslo_concurrency.lockutils [req-181f0a40-7147-4931-a64b-f3a21250edb0 req-cdbe0f54-25a8-46ee-b6fa-7252995194e1 service nova] Releasing lock "refresh_cache-32563e1f-398f-4fc9-866f-d85d9f3f27a7" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1065.841169] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-784c2489-df72-4ab6-b502-a7eab06c668d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.860505] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1065.860505] env[62753]: value = "task-1332760" [ 1065.860505] env[62753]: _type = "Task" [ 1065.860505] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1065.869422] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332760, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1065.913211] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95189f49-2266-42a5-bac3-e659c3cc39ca {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.920209] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64d58b1b-a151-42cd-b4e4-cca1d114a487 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.949555] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "refresh_cache-14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1065.949772] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquired lock "refresh_cache-14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1065.949854] env[62753]: DEBUG nova.network.neutron [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1065.951697] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-426db4a2-f997-4524-96c6-4f67f5d9da5c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.959268] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6e10c7c5-3182-49e9-996a-e870c46ff973 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1065.975487] env[62753]: DEBUG nova.compute.provider_tree [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1066.006269] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52103667-677c-a9c9-5415-e02793a87abf, 'name': SearchDatastore_Task, 'duration_secs': 0.008031} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.006574] env[62753]: DEBUG oslo_concurrency.lockutils [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1066.006806] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1066.007053] env[62753]: DEBUG oslo_concurrency.lockutils [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.007208] env[62753]: DEBUG oslo_concurrency.lockutils [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1066.007392] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1066.007644] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-887bb438-7f3d-4da9-ad33-4aabaae4ad90 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.009463] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1066.015892] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1066.016086] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1066.016774] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e62e05f7-69c0-4535-8f84-1635a75dde35 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.022271] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1066.022271] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52eb0943-8c32-0080-d219-bbc75b828e41" [ 1066.022271] env[62753]: _type = "Task" [ 1066.022271] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.029299] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52eb0943-8c32-0080-d219-bbc75b828e41, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.032791] env[62753]: DEBUG nova.network.neutron [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Successfully created port: ad95cf04-2066-4452-b160-65cf3cddbe87 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1066.212967] env[62753]: DEBUG nova.compute.manager [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1066.284970] env[62753]: DEBUG nova.compute.manager [req-c210a085-7dc3-462b-a6af-927afa2f8358 req-9e3d77a4-08d5-40a7-ad9f-dc78024a2f59 service nova] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Received event network-vif-plugged-3f1e92c2-bb9a-4d5d-a2ea-8f5943d15260 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1066.285271] env[62753]: DEBUG oslo_concurrency.lockutils [req-c210a085-7dc3-462b-a6af-927afa2f8358 req-9e3d77a4-08d5-40a7-ad9f-dc78024a2f59 service nova] Acquiring lock "14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1066.285417] env[62753]: DEBUG oslo_concurrency.lockutils [req-c210a085-7dc3-462b-a6af-927afa2f8358 req-9e3d77a4-08d5-40a7-ad9f-dc78024a2f59 service nova] Lock "14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.285574] env[62753]: DEBUG oslo_concurrency.lockutils [req-c210a085-7dc3-462b-a6af-927afa2f8358 req-9e3d77a4-08d5-40a7-ad9f-dc78024a2f59 service nova] Lock "14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.285716] env[62753]: DEBUG nova.compute.manager [req-c210a085-7dc3-462b-a6af-927afa2f8358 req-9e3d77a4-08d5-40a7-ad9f-dc78024a2f59 service nova] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] No waiting events found dispatching network-vif-plugged-3f1e92c2-bb9a-4d5d-a2ea-8f5943d15260 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1066.285886] env[62753]: WARNING nova.compute.manager [req-c210a085-7dc3-462b-a6af-927afa2f8358 req-9e3d77a4-08d5-40a7-ad9f-dc78024a2f59 service nova] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Received unexpected event network-vif-plugged-3f1e92c2-bb9a-4d5d-a2ea-8f5943d15260 for instance with vm_state building and task_state spawning. [ 1066.286064] env[62753]: DEBUG nova.compute.manager [req-c210a085-7dc3-462b-a6af-927afa2f8358 req-9e3d77a4-08d5-40a7-ad9f-dc78024a2f59 service nova] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Received event network-changed-3f1e92c2-bb9a-4d5d-a2ea-8f5943d15260 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1066.286271] env[62753]: DEBUG nova.compute.manager [req-c210a085-7dc3-462b-a6af-927afa2f8358 req-9e3d77a4-08d5-40a7-ad9f-dc78024a2f59 service nova] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Refreshing instance network info cache due to event network-changed-3f1e92c2-bb9a-4d5d-a2ea-8f5943d15260. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1066.286471] env[62753]: DEBUG oslo_concurrency.lockutils [req-c210a085-7dc3-462b-a6af-927afa2f8358 req-9e3d77a4-08d5-40a7-ad9f-dc78024a2f59 service nova] Acquiring lock "refresh_cache-14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1066.318856] env[62753]: INFO nova.compute.manager [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Took 22.63 seconds to build instance. [ 1066.370916] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332760, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.478438] env[62753]: DEBUG nova.scheduler.client.report [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1066.486820] env[62753]: DEBUG nova.network.neutron [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1066.534738] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52eb0943-8c32-0080-d219-bbc75b828e41, 'name': SearchDatastore_Task, 'duration_secs': 0.007726} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1066.535650] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-2cd21d00-4609-4d19-8c01-3bf8f093ccd1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1066.540888] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1066.540888] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e857e6-bea0-64f0-2f38-126196f153a5" [ 1066.540888] env[62753]: _type = "Task" [ 1066.540888] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1066.548899] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e857e6-bea0-64f0-2f38-126196f153a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.623164] env[62753]: DEBUG nova.network.neutron [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Updating instance_info_cache with network_info: [{"id": "3f1e92c2-bb9a-4d5d-a2ea-8f5943d15260", "address": "fa:16:3e:b6:86:b6", "network": {"id": "e3de96b6-0e52-4885-8747-d3e60da74ef2", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-91543986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b76b178732e41688f767fdd65934193", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f1e92c2-bb", "ovs_interfaceid": "3f1e92c2-bb9a-4d5d-a2ea-8f5943d15260", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1066.821271] env[62753]: DEBUG oslo_concurrency.lockutils [None req-392bd71d-39f5-45c8-a78d-484e953ad68c tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "01d9da0b-f4e5-474c-aab2-a0f4f0c99994" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 24.145s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.872426] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332760, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1066.983442] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.783s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1066.985902] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 7.330s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1066.986214] env[62753]: DEBUG nova.objects.instance [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lazy-loading 'resources' on Instance uuid d7ba8be1-befb-4a0b-9c35-d86e7d586203 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1067.009015] env[62753]: INFO nova.scheduler.client.report [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Deleted allocations for instance 961ced74-10af-4443-857e-545cd890674e [ 1067.015619] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.015619] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.015619] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1067.015619] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62753) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1067.051689] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e857e6-bea0-64f0-2f38-126196f153a5, 'name': SearchDatastore_Task, 'duration_secs': 0.053212} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.051940] env[62753]: DEBUG oslo_concurrency.lockutils [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.052253] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 32563e1f-398f-4fc9-866f-d85d9f3f27a7/32563e1f-398f-4fc9-866f-d85d9f3f27a7.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1067.052539] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-2be934bc-83a4-41f7-8c0c-81b84d4df7c9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.060355] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1067.060355] env[62753]: value = "task-1332761" [ 1067.060355] env[62753]: _type = "Task" [ 1067.060355] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.070071] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332761, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.127017] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Releasing lock "refresh_cache-14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1067.127017] env[62753]: DEBUG nova.compute.manager [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Instance network_info: |[{"id": "3f1e92c2-bb9a-4d5d-a2ea-8f5943d15260", "address": "fa:16:3e:b6:86:b6", "network": {"id": "e3de96b6-0e52-4885-8747-d3e60da74ef2", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-91543986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b76b178732e41688f767fdd65934193", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f1e92c2-bb", "ovs_interfaceid": "3f1e92c2-bb9a-4d5d-a2ea-8f5943d15260", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1067.127017] env[62753]: DEBUG oslo_concurrency.lockutils [req-c210a085-7dc3-462b-a6af-927afa2f8358 req-9e3d77a4-08d5-40a7-ad9f-dc78024a2f59 service nova] Acquired lock "refresh_cache-14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.127017] env[62753]: DEBUG nova.network.neutron [req-c210a085-7dc3-462b-a6af-927afa2f8358 req-9e3d77a4-08d5-40a7-ad9f-dc78024a2f59 service nova] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Refreshing network info cache for port 3f1e92c2-bb9a-4d5d-a2ea-8f5943d15260 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1067.127997] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:b6:86:b6', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '13d625c9-77ec-4edb-a56b-9f37a314cc39', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '3f1e92c2-bb9a-4d5d-a2ea-8f5943d15260', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1067.138616] env[62753]: DEBUG oslo.service.loopingcall [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1067.140148] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1067.140648] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-0bb7832a-5686-46c7-b25d-bec2ad882945 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.161373] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1067.161373] env[62753]: value = "task-1332762" [ 1067.161373] env[62753]: _type = "Task" [ 1067.161373] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.169248] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332762, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.224324] env[62753]: DEBUG nova.compute.manager [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1067.260135] env[62753]: DEBUG nova.virt.hardware [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1067.260418] env[62753]: DEBUG nova.virt.hardware [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1067.260584] env[62753]: DEBUG nova.virt.hardware [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1067.260778] env[62753]: DEBUG nova.virt.hardware [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1067.261166] env[62753]: DEBUG nova.virt.hardware [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1067.261166] env[62753]: DEBUG nova.virt.hardware [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1067.261328] env[62753]: DEBUG nova.virt.hardware [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1067.261467] env[62753]: DEBUG nova.virt.hardware [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1067.261642] env[62753]: DEBUG nova.virt.hardware [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1067.261843] env[62753]: DEBUG nova.virt.hardware [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1067.262054] env[62753]: DEBUG nova.virt.hardware [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1067.262979] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1a83f8e-9fe4-4c1b-948f-d4fe03dbee2c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.271810] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13d70530-715b-46c2-93b7-ccf75966509a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.374190] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332760, 'name': ReconfigVM_Task, 'duration_secs': 1.2186} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.374532] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Reconfigured VM instance instance-00000064 to attach disk [datastore2] 6e417227-895d-4576-b025-7d4a0aafa379/6e417227-895d-4576-b025-7d4a0aafa379.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1067.375334] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b4f8fcae-bb27-45a6-989d-3614bf4965fd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.384250] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1067.384250] env[62753]: value = "task-1332763" [ 1067.384250] env[62753]: _type = "Task" [ 1067.384250] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.394410] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332763, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.521193] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0236f2da-7911-43e6-a71a-15d6610e4325 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "961ced74-10af-4443-857e-545cd890674e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 10.856s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1067.528763] env[62753]: DEBUG nova.compute.manager [req-de5df1b2-3fe0-43c4-86a8-c7be75bdb123 req-aacb7dfb-25ee-4c88-a2aa-a6191f4d866b service nova] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Received event network-changed-d3934283-dc65-4a50-8b00-c466bb372792 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1067.528965] env[62753]: DEBUG nova.compute.manager [req-de5df1b2-3fe0-43c4-86a8-c7be75bdb123 req-aacb7dfb-25ee-4c88-a2aa-a6191f4d866b service nova] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Refreshing instance network info cache due to event network-changed-d3934283-dc65-4a50-8b00-c466bb372792. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1067.529328] env[62753]: DEBUG oslo_concurrency.lockutils [req-de5df1b2-3fe0-43c4-86a8-c7be75bdb123 req-aacb7dfb-25ee-4c88-a2aa-a6191f4d866b service nova] Acquiring lock "refresh_cache-01d9da0b-f4e5-474c-aab2-a0f4f0c99994" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.529620] env[62753]: DEBUG oslo_concurrency.lockutils [req-de5df1b2-3fe0-43c4-86a8-c7be75bdb123 req-aacb7dfb-25ee-4c88-a2aa-a6191f4d866b service nova] Acquired lock "refresh_cache-01d9da0b-f4e5-474c-aab2-a0f4f0c99994" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.529924] env[62753]: DEBUG nova.network.neutron [req-de5df1b2-3fe0-43c4-86a8-c7be75bdb123 req-aacb7dfb-25ee-4c88-a2aa-a6191f4d866b service nova] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Refreshing network info cache for port d3934283-dc65-4a50-8b00-c466bb372792 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1067.579688] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332761, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.484185} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.582187] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 32563e1f-398f-4fc9-866f-d85d9f3f27a7/32563e1f-398f-4fc9-866f-d85d9f3f27a7.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1067.582418] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1067.583666] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-ed1c4fee-25b5-489d-850a-a68b7dabc218 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.589124] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1067.589124] env[62753]: value = "task-1332764" [ 1067.589124] env[62753]: _type = "Task" [ 1067.589124] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.598166] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332764, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.676331] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332762, 'name': CreateVM_Task, 'duration_secs': 0.454394} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.676502] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1067.677421] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1067.677595] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1067.678062] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1067.678342] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-12cbc9ea-a6e6-4de2-a3c2-5c1a2b90f9d9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.683462] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1067.683462] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52285fcc-22c4-1054-37ac-87216801bf9d" [ 1067.683462] env[62753]: _type = "Task" [ 1067.683462] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.688943] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97ae1a5e-b096-4dfa-843b-17840b15d1ea {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.694160] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52285fcc-22c4-1054-37ac-87216801bf9d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1067.698275] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f217be8d-c1b8-45ac-a5da-a4253bebbff3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.730241] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-172f9df4-929f-4e02-9318-b006fa428d89 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.737685] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-92e8818d-7116-4e14-b9a0-11b2f90e6027 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.750371] env[62753]: DEBUG nova.compute.provider_tree [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1067.791471] env[62753]: DEBUG nova.network.neutron [req-c210a085-7dc3-462b-a6af-927afa2f8358 req-9e3d77a4-08d5-40a7-ad9f-dc78024a2f59 service nova] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Updated VIF entry in instance network info cache for port 3f1e92c2-bb9a-4d5d-a2ea-8f5943d15260. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1067.791827] env[62753]: DEBUG nova.network.neutron [req-c210a085-7dc3-462b-a6af-927afa2f8358 req-9e3d77a4-08d5-40a7-ad9f-dc78024a2f59 service nova] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Updating instance_info_cache with network_info: [{"id": "3f1e92c2-bb9a-4d5d-a2ea-8f5943d15260", "address": "fa:16:3e:b6:86:b6", "network": {"id": "e3de96b6-0e52-4885-8747-d3e60da74ef2", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-91543986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b76b178732e41688f767fdd65934193", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap3f1e92c2-bb", "ovs_interfaceid": "3f1e92c2-bb9a-4d5d-a2ea-8f5943d15260", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1067.842274] env[62753]: DEBUG nova.network.neutron [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Successfully updated port: ad95cf04-2066-4452-b160-65cf3cddbe87 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1067.894781] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332763, 'name': Rename_Task, 'duration_secs': 0.220925} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1067.895090] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1067.895378] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4dc312ab-27b2-4db6-8d13-233c4604e39d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1067.901052] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1067.901052] env[62753]: value = "task-1332765" [ 1067.901052] env[62753]: _type = "Task" [ 1067.901052] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1067.908281] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332765, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.011068] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1068.100435] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332764, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071254} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.100708] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1068.101498] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3b0c76af-1da5-47fe-96cc-340ef8db1f7d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.128055] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Reconfiguring VM instance instance-00000065 to attach disk [datastore2] 32563e1f-398f-4fc9-866f-d85d9f3f27a7/32563e1f-398f-4fc9-866f-d85d9f3f27a7.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1068.128392] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-1c75dc00-60df-4b12-9c0f-f14982ba0260 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.149874] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1068.149874] env[62753]: value = "task-1332766" [ 1068.149874] env[62753]: _type = "Task" [ 1068.149874] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.157870] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332766, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.194826] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52285fcc-22c4-1054-37ac-87216801bf9d, 'name': SearchDatastore_Task, 'duration_secs': 0.009953} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.197507] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.197771] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1068.198036] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.198203] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.198383] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1068.198649] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-8f9ba23e-ab8e-4cf9-b71e-75067c1300cc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.206200] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1068.206386] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1068.207102] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ddb8f999-845e-4b93-85ff-9ea55f4e39f4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.211786] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1068.211786] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52295f52-90f2-c6fc-9b37-386e0b640e5a" [ 1068.211786] env[62753]: _type = "Task" [ 1068.211786] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.219014] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52295f52-90f2-c6fc-9b37-386e0b640e5a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.254055] env[62753]: DEBUG nova.scheduler.client.report [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1068.267795] env[62753]: DEBUG nova.network.neutron [req-de5df1b2-3fe0-43c4-86a8-c7be75bdb123 req-aacb7dfb-25ee-4c88-a2aa-a6191f4d866b service nova] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Updated VIF entry in instance network info cache for port d3934283-dc65-4a50-8b00-c466bb372792. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1068.268180] env[62753]: DEBUG nova.network.neutron [req-de5df1b2-3fe0-43c4-86a8-c7be75bdb123 req-aacb7dfb-25ee-4c88-a2aa-a6191f4d866b service nova] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Updating instance_info_cache with network_info: [{"id": "d3934283-dc65-4a50-8b00-c466bb372792", "address": "fa:16:3e:bf:92:e9", "network": {"id": "61c6e640-d2e6-4b3b-b191-7da9249dd339", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-703836479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.4", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.129", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8925b84dcf9a47fbaf2eb8044b3850fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapd3934283-dc", "ovs_interfaceid": "d3934283-dc65-4a50-8b00-c466bb372792", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1068.294271] env[62753]: DEBUG oslo_concurrency.lockutils [req-c210a085-7dc3-462b-a6af-927afa2f8358 req-9e3d77a4-08d5-40a7-ad9f-dc78024a2f59 service nova] Releasing lock "refresh_cache-14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.313117] env[62753]: DEBUG nova.compute.manager [req-25a3f07c-d0a0-4b9f-a30b-96c187a50a54 req-1c408953-a854-4539-9b17-cf61acace251 service nova] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Received event network-vif-plugged-ad95cf04-2066-4452-b160-65cf3cddbe87 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1068.313398] env[62753]: DEBUG oslo_concurrency.lockutils [req-25a3f07c-d0a0-4b9f-a30b-96c187a50a54 req-1c408953-a854-4539-9b17-cf61acace251 service nova] Acquiring lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1068.313598] env[62753]: DEBUG oslo_concurrency.lockutils [req-25a3f07c-d0a0-4b9f-a30b-96c187a50a54 req-1c408953-a854-4539-9b17-cf61acace251 service nova] Lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.313825] env[62753]: DEBUG oslo_concurrency.lockutils [req-25a3f07c-d0a0-4b9f-a30b-96c187a50a54 req-1c408953-a854-4539-9b17-cf61acace251 service nova] Lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.313965] env[62753]: DEBUG nova.compute.manager [req-25a3f07c-d0a0-4b9f-a30b-96c187a50a54 req-1c408953-a854-4539-9b17-cf61acace251 service nova] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] No waiting events found dispatching network-vif-plugged-ad95cf04-2066-4452-b160-65cf3cddbe87 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1068.314163] env[62753]: WARNING nova.compute.manager [req-25a3f07c-d0a0-4b9f-a30b-96c187a50a54 req-1c408953-a854-4539-9b17-cf61acace251 service nova] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Received unexpected event network-vif-plugged-ad95cf04-2066-4452-b160-65cf3cddbe87 for instance with vm_state building and task_state spawning. [ 1068.314351] env[62753]: DEBUG nova.compute.manager [req-25a3f07c-d0a0-4b9f-a30b-96c187a50a54 req-1c408953-a854-4539-9b17-cf61acace251 service nova] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Received event network-changed-ad95cf04-2066-4452-b160-65cf3cddbe87 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1068.314515] env[62753]: DEBUG nova.compute.manager [req-25a3f07c-d0a0-4b9f-a30b-96c187a50a54 req-1c408953-a854-4539-9b17-cf61acace251 service nova] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Refreshing instance network info cache due to event network-changed-ad95cf04-2066-4452-b160-65cf3cddbe87. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1068.314756] env[62753]: DEBUG oslo_concurrency.lockutils [req-25a3f07c-d0a0-4b9f-a30b-96c187a50a54 req-1c408953-a854-4539-9b17-cf61acace251 service nova] Acquiring lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.314910] env[62753]: DEBUG oslo_concurrency.lockutils [req-25a3f07c-d0a0-4b9f-a30b-96c187a50a54 req-1c408953-a854-4539-9b17-cf61acace251 service nova] Acquired lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1068.315113] env[62753]: DEBUG nova.network.neutron [req-25a3f07c-d0a0-4b9f-a30b-96c187a50a54 req-1c408953-a854-4539-9b17-cf61acace251 service nova] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Refreshing network info cache for port ad95cf04-2066-4452-b160-65cf3cddbe87 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1068.345215] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1068.411332] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332765, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.659745] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332766, 'name': ReconfigVM_Task, 'duration_secs': 0.262277} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.660016] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Reconfigured VM instance instance-00000065 to attach disk [datastore2] 32563e1f-398f-4fc9-866f-d85d9f3f27a7/32563e1f-398f-4fc9-866f-d85d9f3f27a7.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1068.660714] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-a4a7e98e-4a69-44d9-8306-458b92cfc1d9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.666269] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1068.666269] env[62753]: value = "task-1332767" [ 1068.666269] env[62753]: _type = "Task" [ 1068.666269] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.674690] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332767, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.721645] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52295f52-90f2-c6fc-9b37-386e0b640e5a, 'name': SearchDatastore_Task, 'duration_secs': 0.008142} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1068.722464] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c63c06ec-f9e3-4dfa-ade0-9ed2c6d41146 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1068.727698] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1068.727698] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a3104a-d8bb-6206-2355-5c04d4b01044" [ 1068.727698] env[62753]: _type = "Task" [ 1068.727698] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1068.735046] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a3104a-d8bb-6206-2355-5c04d4b01044, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.759130] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.773s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1068.761477] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 8.822s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1068.771492] env[62753]: DEBUG oslo_concurrency.lockutils [req-de5df1b2-3fe0-43c4-86a8-c7be75bdb123 req-aacb7dfb-25ee-4c88-a2aa-a6191f4d866b service nova] Releasing lock "refresh_cache-01d9da0b-f4e5-474c-aab2-a0f4f0c99994" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1068.784411] env[62753]: INFO nova.scheduler.client.report [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Deleted allocations for instance d7ba8be1-befb-4a0b-9c35-d86e7d586203 [ 1068.847895] env[62753]: DEBUG nova.network.neutron [req-25a3f07c-d0a0-4b9f-a30b-96c187a50a54 req-1c408953-a854-4539-9b17-cf61acace251 service nova] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1068.912502] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332765, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1068.925837] env[62753]: DEBUG nova.network.neutron [req-25a3f07c-d0a0-4b9f-a30b-96c187a50a54 req-1c408953-a854-4539-9b17-cf61acace251 service nova] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1069.176367] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332767, 'name': Rename_Task, 'duration_secs': 0.142491} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.177052] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1069.177311] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-69e052d8-17c7-410c-ba46-4b8a2554f8c5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.183684] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1069.183684] env[62753]: value = "task-1332768" [ 1069.183684] env[62753]: _type = "Task" [ 1069.183684] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.190812] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332768, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.236975] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a3104a-d8bb-6206-2355-5c04d4b01044, 'name': SearchDatastore_Task, 'duration_secs': 0.008912} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.237404] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.237670] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac/14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1069.237928] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-96c9fc39-453c-4f25-a2a0-c1e174da7a49 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.243374] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1069.243374] env[62753]: value = "task-1332769" [ 1069.243374] env[62753]: _type = "Task" [ 1069.243374] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1069.250878] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332769, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.266780] env[62753]: INFO nova.compute.claims [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1069.292771] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5d9594aa-95a0-4935-b7b3-d9a9e2fb7ea2 tempest-ListServersNegativeTestJSON-1492396842 tempest-ListServersNegativeTestJSON-1492396842-project-member] Lock "d7ba8be1-befb-4a0b-9c35-d86e7d586203" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 12.701s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1069.413680] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332765, 'name': PowerOnVM_Task} progress is 71%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.428226] env[62753]: DEBUG oslo_concurrency.lockutils [req-25a3f07c-d0a0-4b9f-a30b-96c187a50a54 req-1c408953-a854-4539-9b17-cf61acace251 service nova] Releasing lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1069.428669] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1069.428845] env[62753]: DEBUG nova.network.neutron [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1069.694857] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332768, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.753414] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332769, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1069.773409] env[62753]: INFO nova.compute.resource_tracker [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Updating resource usage from migration 6a1e13cb-e250-4311-b57b-51a150299b67 [ 1069.917748] env[62753]: DEBUG oslo_vmware.api [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332765, 'name': PowerOnVM_Task, 'duration_secs': 1.847163} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1069.918963] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1069.918963] env[62753]: INFO nova.compute.manager [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Took 9.43 seconds to spawn the instance on the hypervisor. [ 1069.918963] env[62753]: DEBUG nova.compute.manager [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1069.919571] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-48e165a4-827b-4583-b010-2520229a9d53 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1069.964963] env[62753]: DEBUG nova.network.neutron [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1070.012517] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1070.016644] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a08c0e76-56cc-4046-a160-6cc7d2598b4e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.025850] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-958919ea-af39-454b-9c22-b89d5138ea03 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.063069] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c8c686c6-8e8d-4d10-b4f4-5c65c8a82aad {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.073736] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5a4682c8-99cc-4f52-bcec-f120a1a94c91 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.089762] env[62753]: DEBUG nova.compute.provider_tree [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1070.144536] env[62753]: DEBUG nova.network.neutron [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance_info_cache with network_info: [{"id": "ad95cf04-2066-4452-b160-65cf3cddbe87", "address": "fa:16:3e:eb:5e:01", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad95cf04-20", "ovs_interfaceid": "ad95cf04-2066-4452-b160-65cf3cddbe87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1070.197438] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332768, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.254196] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332769, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.982627} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.254532] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac/14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1070.254825] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1070.255126] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-63a5703a-8828-4065-a520-84b4048d422c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.261376] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1070.261376] env[62753]: value = "task-1332770" [ 1070.261376] env[62753]: _type = "Task" [ 1070.261376] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.268852] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332770, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.444735] env[62753]: INFO nova.compute.manager [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Took 15.85 seconds to build instance. [ 1070.593802] env[62753]: DEBUG nova.scheduler.client.report [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1070.647738] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1070.648096] env[62753]: DEBUG nova.compute.manager [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Instance network_info: |[{"id": "ad95cf04-2066-4452-b160-65cf3cddbe87", "address": "fa:16:3e:eb:5e:01", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad95cf04-20", "ovs_interfaceid": "ad95cf04-2066-4452-b160-65cf3cddbe87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1070.648963] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:eb:5e:01', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'ad95cf04-2066-4452-b160-65cf3cddbe87', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1070.656620] env[62753]: DEBUG oslo.service.loopingcall [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1070.656834] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1070.657075] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-e9aecbfc-3165-4127-bbff-29d96ac5e90b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.677387] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1070.677387] env[62753]: value = "task-1332771" [ 1070.677387] env[62753]: _type = "Task" [ 1070.677387] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.687113] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332771, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.694470] env[62753]: DEBUG oslo_vmware.api [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332768, 'name': PowerOnVM_Task, 'duration_secs': 1.303062} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.694729] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1070.694943] env[62753]: INFO nova.compute.manager [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Took 7.98 seconds to spawn the instance on the hypervisor. [ 1070.695143] env[62753]: DEBUG nova.compute.manager [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1070.695910] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03c2b948-8a8d-4d65-ae71-30832969e915 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.772232] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332770, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.074643} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1070.772232] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1070.772619] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abfccb35-b01d-49d9-9fa8-2e36093a45df {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.797375] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Reconfiguring VM instance instance-00000066 to attach disk [datastore2] 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac/14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1070.797710] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-706003c2-b61d-4a6e-b70d-599019aba0cc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1070.818253] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1070.818253] env[62753]: value = "task-1332772" [ 1070.818253] env[62753]: _type = "Task" [ 1070.818253] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1070.829298] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332772, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1070.946315] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4487d0ba-d6b7-40f8-a443-14009b9477f5 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "6e417227-895d-4576-b025-7d4a0aafa379" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.363s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.006526] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.010363] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1071.100032] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.338s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.100032] env[62753]: INFO nova.compute.manager [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Migrating [ 1071.107724] env[62753]: DEBUG oslo_concurrency.lockutils [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 6.434s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1071.109282] env[62753]: INFO nova.compute.claims [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1071.187383] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332771, 'name': CreateVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.213870] env[62753]: INFO nova.compute.manager [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Took 16.30 seconds to build instance. [ 1071.327816] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332772, 'name': ReconfigVM_Task, 'duration_secs': 0.257574} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.328210] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Reconfigured VM instance instance-00000066 to attach disk [datastore2] 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac/14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1071.328935] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-85acbe61-e55f-4ef9-a050-1c5ae031fa99 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.335323] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1071.335323] env[62753]: value = "task-1332773" [ 1071.335323] env[62753]: _type = "Task" [ 1071.335323] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.344065] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332773, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.513958] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1071.621323] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "refresh_cache-eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.621323] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "refresh_cache-eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.621490] env[62753]: DEBUG nova.network.neutron [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1071.689659] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332771, 'name': CreateVM_Task, 'duration_secs': 0.684462} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.689912] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1071.690670] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1071.690915] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1071.691324] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1071.691635] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-35cd967d-e0aa-489c-acbd-4980dd9a65d2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.696464] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1071.696464] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]526b8c47-4126-232e-3ff4-da8a33524abb" [ 1071.696464] env[62753]: _type = "Task" [ 1071.696464] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.704187] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]526b8c47-4126-232e-3ff4-da8a33524abb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1071.716802] env[62753]: DEBUG oslo_concurrency.lockutils [None req-767e619b-8e44-44ca-8a98-9cc3dd200ed4 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "32563e1f-398f-4fc9-866f-d85d9f3f27a7" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 17.808s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1071.845873] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332773, 'name': Rename_Task, 'duration_secs': 0.133459} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1071.846246] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1071.846519] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd5a2350-a450-4ca0-9691-c28c94df36e2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1071.852875] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1071.852875] env[62753]: value = "task-1332774" [ 1071.852875] env[62753]: _type = "Task" [ 1071.852875] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1071.860329] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332774, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.206967] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]526b8c47-4126-232e-3ff4-da8a33524abb, 'name': SearchDatastore_Task, 'duration_secs': 0.026511} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.209557] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1072.209774] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1072.210046] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1072.210212] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1072.210396] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1072.210906] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-ebbd3b06-7d5b-4c30-a9e8-b2e300c99c90 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.223897] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1072.224208] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1072.227519] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4803bb0e-aa83-4406-acb9-5fd7028bf5b9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.233487] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1072.233487] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5221faf3-0a1a-14ea-8d80-1fa16c6b7166" [ 1072.233487] env[62753]: _type = "Task" [ 1072.233487] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.241755] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5221faf3-0a1a-14ea-8d80-1fa16c6b7166, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.270236] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca13a04-6c66-4116-bd8d-0f6e57692847 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.278404] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0e0e3863-6e0c-4dd2-9a8d-dba92a7c3e87 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.314577] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62f51fe9-a511-409a-abc7-97705d5588f9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.322843] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b78916f8-2e37-48a5-a411-b9819272dc7a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.338940] env[62753]: DEBUG nova.compute.provider_tree [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1072.371159] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332774, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.427633] env[62753]: DEBUG nova.network.neutron [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Updating instance_info_cache with network_info: [{"id": "93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9", "address": "fa:16:3e:90:68:e2", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93d26d06-fe", "ovs_interfaceid": "93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1072.743618] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5221faf3-0a1a-14ea-8d80-1fa16c6b7166, 'name': SearchDatastore_Task, 'duration_secs': 0.010471} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1072.744496] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-acf9bfea-5cd0-4815-a712-dfdf377b3827 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1072.751106] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1072.751106] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]526e200e-da15-2b37-915f-09afa35e0d00" [ 1072.751106] env[62753]: _type = "Task" [ 1072.751106] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1072.758610] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]526e200e-da15-2b37-915f-09afa35e0d00, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.842852] env[62753]: DEBUG nova.scheduler.client.report [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1072.864725] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332774, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1072.933647] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "refresh_cache-eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.262226] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]526e200e-da15-2b37-915f-09afa35e0d00, 'name': SearchDatastore_Task, 'duration_secs': 0.038651} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.262590] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1073.262907] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] ca01cd3f-7f0e-4c61-a704-74f9945792b4/ca01cd3f-7f0e-4c61-a704-74f9945792b4.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1073.263253] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-01cc219a-17af-4c4d-a3e0-357c8ca0e9c2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.270189] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1073.270189] env[62753]: value = "task-1332775" [ 1073.270189] env[62753]: _type = "Task" [ 1073.270189] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1073.279079] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332775, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.350373] env[62753]: DEBUG oslo_concurrency.lockutils [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.243s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.351229] env[62753]: DEBUG nova.compute.manager [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1073.354658] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.841s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.354937] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1073.355181] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62753) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1073.356424] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2dc0032a-63df-402e-a522-163153fc8bbe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.371448] env[62753]: DEBUG oslo_vmware.api [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332774, 'name': PowerOnVM_Task, 'duration_secs': 1.268645} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1073.371885] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1073.372333] env[62753]: INFO nova.compute.manager [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Took 8.40 seconds to spawn the instance on the hypervisor. [ 1073.372333] env[62753]: DEBUG nova.compute.manager [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1073.373641] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41fe0bc0-23ca-445f-ad02-528ddbb72041 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.379071] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d132a61d-6f0b-449c-9e9f-da21706a67a0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.401723] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-387c0de2-2a68-4a4b-8d7e-c4433f2fc59c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.409315] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-159e1dfd-2057-4836-bd70-310407b77201 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1073.446907] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180648MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=62753) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1073.447109] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1073.447422] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1073.782025] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332775, 'name': CopyVirtualDisk_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1073.858405] env[62753]: DEBUG nova.compute.utils [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1073.859874] env[62753]: DEBUG nova.compute.manager [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1073.860107] env[62753]: DEBUG nova.network.neutron [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1073.908679] env[62753]: DEBUG nova.policy [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd65bf0df728640ef9104e14d6320ed01', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '891d1efbcac34ceba5fa72ea6b5fe38b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1073.916401] env[62753]: INFO nova.compute.manager [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Took 18.68 seconds to build instance. [ 1074.235007] env[62753]: DEBUG nova.network.neutron [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Successfully created port: 25e4c4c8-9d78-4a88-9a5e-4a364ee574a5 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1074.287449] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332775, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.656115} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.287449] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] ca01cd3f-7f0e-4c61-a704-74f9945792b4/ca01cd3f-7f0e-4c61-a704-74f9945792b4.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1074.287449] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1074.287449] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-2e970d9c-cb63-491e-803d-904481370f20 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.296864] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1074.296864] env[62753]: value = "task-1332776" [ 1074.296864] env[62753]: _type = "Task" [ 1074.296864] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.305027] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332776, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1074.364059] env[62753]: DEBUG nova.compute.manager [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1074.416460] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd2ade5a-af68-4abe-89ba-33a595a66993 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 20.194s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1074.461388] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Applying migration context for instance eb86a978-27fa-41b2-a7e5-4b3a8ba0a152 as it has an incoming, in-progress migration 6a1e13cb-e250-4311-b57b-51a150299b67. Migration status is migrating {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1074.465197] env[62753]: INFO nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Updating resource usage from migration 6a1e13cb-e250-4311-b57b-51a150299b67 [ 1074.467192] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6162e911-76fd-4dc8-89da-257070a7a7a2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.499243] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Updating instance 'eb86a978-27fa-41b2-a7e5-4b3a8ba0a152' progress to 0 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1074.503598] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 04fd0022-f20a-4217-9e47-5381635b17a6 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.503877] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 01d9da0b-f4e5-474c-aab2-a0f4f0c99994 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.504133] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 6e417227-895d-4576-b025-7d4a0aafa379 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.504367] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 32563e1f-398f-4fc9-866f-d85d9f3f27a7 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.504564] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.504780] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance ca01cd3f-7f0e-4c61-a704-74f9945792b4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.504989] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Migration 6a1e13cb-e250-4311-b57b-51a150299b67 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1074.505568] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance eb86a978-27fa-41b2-a7e5-4b3a8ba0a152 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.505823] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 22cea165-cfde-403f-a0e2-82a86f5baa51 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1074.507019] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Total usable vcpus: 48, total allocated vcpus: 9 {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1074.507019] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2368MB phys_disk=200GB used_disk=9GB total_vcpus=48 used_vcpus=9 pci_stats=[] {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1074.639297] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da7bc09e-162b-47e0-b915-be88f8f3087e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.647880] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2b5287f3-7100-49a8-b8a6-c2d161a57bb3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.679655] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-06462f38-5c4e-47b8-bebe-d67600438663 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.688058] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c84cc9b1-21d1-411f-a206-64d27f410bc9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.700522] env[62753]: DEBUG nova.compute.provider_tree [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1074.807172] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332776, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.188439} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1074.808251] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1074.808300] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7aa631e-2a31-4806-afb6-e5fa542d0fea {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.830798] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] ca01cd3f-7f0e-4c61-a704-74f9945792b4/ca01cd3f-7f0e-4c61-a704-74f9945792b4.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1074.831028] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ca973bf3-f7a9-4a4b-80a0-c679d75a0101 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1074.851174] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1074.851174] env[62753]: value = "task-1332777" [ 1074.851174] env[62753]: _type = "Task" [ 1074.851174] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1074.861455] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332777, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.011533] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1075.011869] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-20d230d4-f51f-4d86-9569-d981d8c2fa07 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.019524] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1075.019524] env[62753]: value = "task-1332778" [ 1075.019524] env[62753]: _type = "Task" [ 1075.019524] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1075.028845] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332778, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.132821] env[62753]: DEBUG oslo_concurrency.lockutils [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "a2e7fa98-566e-40f1-8da7-9318de89c14c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.132821] env[62753]: DEBUG oslo_concurrency.lockutils [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "a2e7fa98-566e-40f1-8da7-9318de89c14c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.203791] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1075.360567] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332777, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.375330] env[62753]: DEBUG nova.compute.manager [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1075.404201] env[62753]: DEBUG nova.virt.hardware [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1075.404849] env[62753]: DEBUG nova.virt.hardware [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1075.404849] env[62753]: DEBUG nova.virt.hardware [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1075.404974] env[62753]: DEBUG nova.virt.hardware [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1075.407085] env[62753]: DEBUG nova.virt.hardware [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1075.407085] env[62753]: DEBUG nova.virt.hardware [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1075.407085] env[62753]: DEBUG nova.virt.hardware [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1075.407085] env[62753]: DEBUG nova.virt.hardware [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1075.407085] env[62753]: DEBUG nova.virt.hardware [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1075.407085] env[62753]: DEBUG nova.virt.hardware [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1075.407085] env[62753]: DEBUG nova.virt.hardware [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1075.407463] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e3c1c1a3-30a5-4518-b828-67154e723a77 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.415360] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87be11d7-898c-4aef-9a44-962c29f712dd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.515274] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "51790f51-c464-48cc-9891-41cae5a8f63e" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.516363] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "51790f51-c464-48cc-9891-41cae5a8f63e" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.529464] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332778, 'name': PowerOffVM_Task, 'duration_secs': 0.459913} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1075.529715] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1075.529918] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Updating instance 'eb86a978-27fa-41b2-a7e5-4b3a8ba0a152' progress to 17 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1075.634574] env[62753]: DEBUG nova.compute.manager [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1075.660176] env[62753]: DEBUG nova.compute.manager [req-ec39bd6e-f42e-4f85-8427-f05ffc405bb8 req-e57fbfbd-711b-453c-969f-31b4408c857a service nova] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Received event network-vif-plugged-25e4c4c8-9d78-4a88-9a5e-4a364ee574a5 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1075.660422] env[62753]: DEBUG oslo_concurrency.lockutils [req-ec39bd6e-f42e-4f85-8427-f05ffc405bb8 req-e57fbfbd-711b-453c-969f-31b4408c857a service nova] Acquiring lock "22cea165-cfde-403f-a0e2-82a86f5baa51-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.660649] env[62753]: DEBUG oslo_concurrency.lockutils [req-ec39bd6e-f42e-4f85-8427-f05ffc405bb8 req-e57fbfbd-711b-453c-969f-31b4408c857a service nova] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.660825] env[62753]: DEBUG oslo_concurrency.lockutils [req-ec39bd6e-f42e-4f85-8427-f05ffc405bb8 req-e57fbfbd-711b-453c-969f-31b4408c857a service nova] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.661055] env[62753]: DEBUG nova.compute.manager [req-ec39bd6e-f42e-4f85-8427-f05ffc405bb8 req-e57fbfbd-711b-453c-969f-31b4408c857a service nova] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] No waiting events found dispatching network-vif-plugged-25e4c4c8-9d78-4a88-9a5e-4a364ee574a5 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1075.664051] env[62753]: WARNING nova.compute.manager [req-ec39bd6e-f42e-4f85-8427-f05ffc405bb8 req-e57fbfbd-711b-453c-969f-31b4408c857a service nova] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Received unexpected event network-vif-plugged-25e4c4c8-9d78-4a88-9a5e-4a364ee574a5 for instance with vm_state building and task_state spawning. [ 1075.709342] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1075.709618] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.262s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1075.753132] env[62753]: DEBUG nova.network.neutron [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Successfully updated port: 25e4c4c8-9d78-4a88-9a5e-4a364ee574a5 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1075.863489] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332777, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1075.953382] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e6ca60ef-1101-413d-8b08-e1c19f739338 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "6e417227-895d-4576-b025-7d4a0aafa379" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1075.953671] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e6ca60ef-1101-413d-8b08-e1c19f739338 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "6e417227-895d-4576-b025-7d4a0aafa379" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1075.953862] env[62753]: DEBUG nova.compute.manager [None req-e6ca60ef-1101-413d-8b08-e1c19f739338 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1075.954866] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d4910788-ebc3-4ee0-9823-89e389987747 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1075.963047] env[62753]: DEBUG nova.compute.manager [None req-e6ca60ef-1101-413d-8b08-e1c19f739338 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62753) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1075.963047] env[62753]: DEBUG nova.objects.instance [None req-e6ca60ef-1101-413d-8b08-e1c19f739338 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lazy-loading 'flavor' on Instance uuid 6e417227-895d-4576-b025-7d4a0aafa379 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1076.021017] env[62753]: DEBUG nova.compute.manager [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1076.037253] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1076.037521] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1076.037694] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1076.037883] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1076.038362] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1076.038671] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1076.038962] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1076.039410] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1076.040237] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1076.040493] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1076.040744] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1076.048435] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5b0c9ad3-7117-47ba-9196-fd42aeea0c3d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.068171] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1076.068171] env[62753]: value = "task-1332779" [ 1076.068171] env[62753]: _type = "Task" [ 1076.068171] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.077350] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332779, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.157345] env[62753]: DEBUG oslo_concurrency.lockutils [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.157614] env[62753]: DEBUG oslo_concurrency.lockutils [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1076.159153] env[62753]: INFO nova.compute.claims [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1076.259889] env[62753]: DEBUG oslo_concurrency.lockutils [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1076.260687] env[62753]: DEBUG oslo_concurrency.lockutils [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1076.260687] env[62753]: DEBUG nova.network.neutron [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1076.364457] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332777, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.468862] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6ca60ef-1101-413d-8b08-e1c19f739338 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1076.469159] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-58a9e0c1-1be3-49b1-86c4-4f5cf0f6786d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.477230] env[62753]: DEBUG oslo_vmware.api [None req-e6ca60ef-1101-413d-8b08-e1c19f739338 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1076.477230] env[62753]: value = "task-1332780" [ 1076.477230] env[62753]: _type = "Task" [ 1076.477230] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.485584] env[62753]: DEBUG oslo_vmware.api [None req-e6ca60ef-1101-413d-8b08-e1c19f739338 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332780, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.539017] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1076.578781] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332779, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.792780] env[62753]: DEBUG nova.network.neutron [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1076.862391] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332777, 'name': ReconfigVM_Task, 'duration_secs': 1.835676} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.862690] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Reconfigured VM instance instance-00000067 to attach disk [datastore1] ca01cd3f-7f0e-4c61-a704-74f9945792b4/ca01cd3f-7f0e-4c61-a704-74f9945792b4.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1076.865709] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-15c5f719-f093-435b-a20a-115912e163df {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1076.872094] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1076.872094] env[62753]: value = "task-1332781" [ 1076.872094] env[62753]: _type = "Task" [ 1076.872094] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1076.880874] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332781, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1076.938719] env[62753]: DEBUG nova.network.neutron [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance_info_cache with network_info: [{"id": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "address": "fa:16:3e:88:6a:1d", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25e4c4c8-9d", "ovs_interfaceid": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1076.987951] env[62753]: DEBUG oslo_vmware.api [None req-e6ca60ef-1101-413d-8b08-e1c19f739338 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332780, 'name': PowerOffVM_Task, 'duration_secs': 0.339434} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1076.988254] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-e6ca60ef-1101-413d-8b08-e1c19f739338 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1076.988437] env[62753]: DEBUG nova.compute.manager [None req-e6ca60ef-1101-413d-8b08-e1c19f739338 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1076.989255] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-521c8c50-a8bd-4cf8-a4cb-07b44088b94e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.079217] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332779, 'name': ReconfigVM_Task, 'duration_secs': 0.702075} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.079556] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Updating instance 'eb86a978-27fa-41b2-a7e5-4b3a8ba0a152' progress to 33 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1077.310059] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c4344e0-addc-4396-aa18-52fe180cc911 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.318528] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-59e558ba-58db-47f9-97ce-6f45702f92e8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.349944] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55ca4dfc-cf85-43d0-bda8-518a5dca4dfa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.357364] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edb18567-8463-4011-a809-085fe159390d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.371412] env[62753]: DEBUG nova.compute.provider_tree [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1077.380995] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332781, 'name': Rename_Task, 'duration_secs': 0.48123} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.381276] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1077.381489] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-9b69e724-5abe-43f6-84fc-49aa607a8b74 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.388513] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1077.388513] env[62753]: value = "task-1332782" [ 1077.388513] env[62753]: _type = "Task" [ 1077.388513] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.395878] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332782, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.441396] env[62753]: DEBUG oslo_concurrency.lockutils [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1077.441728] env[62753]: DEBUG nova.compute.manager [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Instance network_info: |[{"id": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "address": "fa:16:3e:88:6a:1d", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25e4c4c8-9d", "ovs_interfaceid": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1077.442286] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:88:6a:1d', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cae1d6a8-cbba-4bbf-af10-ba5467340475', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '25e4c4c8-9d78-4a88-9a5e-4a364ee574a5', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1077.449886] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Creating folder: Project (891d1efbcac34ceba5fa72ea6b5fe38b). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1077.450478] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-daa281bb-31c8-4282-a2c8-29fbe91f1c63 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.461246] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Created folder: Project (891d1efbcac34ceba5fa72ea6b5fe38b) in parent group-v284541. [ 1077.461424] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Creating folder: Instances. Parent ref: group-v284682. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1077.461651] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-0d22bc0e-7935-4bf4-9d34-d6ac88d823cf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.469145] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Created folder: Instances in parent group-v284682. [ 1077.469378] env[62753]: DEBUG oslo.service.loopingcall [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1077.469565] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1077.469761] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-cd203381-39c4-46d4-b95f-509979036e33 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.487605] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1077.487605] env[62753]: value = "task-1332785" [ 1077.487605] env[62753]: _type = "Task" [ 1077.487605] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.494768] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332785, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.500705] env[62753]: DEBUG oslo_concurrency.lockutils [None req-e6ca60ef-1101-413d-8b08-e1c19f739338 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "6e417227-895d-4576-b025-7d4a0aafa379" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.547s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1077.586719] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1077.586989] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1077.587181] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1077.587391] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1077.587548] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1077.587702] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1077.587915] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1077.588098] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1077.588274] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1077.588446] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1077.588624] env[62753]: DEBUG nova.virt.hardware [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1077.593798] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Reconfiguring VM instance instance-00000062 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1077.594128] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b85202a-1105-4d9a-8a13-8bc879b71226 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1077.611978] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1077.611978] env[62753]: value = "task-1332786" [ 1077.611978] env[62753]: _type = "Task" [ 1077.611978] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1077.619791] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332786, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.687470] env[62753]: DEBUG nova.compute.manager [req-c0a8ee18-a1a2-4213-b5f0-c62323d38e24 req-35644b6e-5f5a-4080-be01-cefa2a813dc4 service nova] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Received event network-changed-25e4c4c8-9d78-4a88-9a5e-4a364ee574a5 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1077.687800] env[62753]: DEBUG nova.compute.manager [req-c0a8ee18-a1a2-4213-b5f0-c62323d38e24 req-35644b6e-5f5a-4080-be01-cefa2a813dc4 service nova] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Refreshing instance network info cache due to event network-changed-25e4c4c8-9d78-4a88-9a5e-4a364ee574a5. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1077.688069] env[62753]: DEBUG oslo_concurrency.lockutils [req-c0a8ee18-a1a2-4213-b5f0-c62323d38e24 req-35644b6e-5f5a-4080-be01-cefa2a813dc4 service nova] Acquiring lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.688258] env[62753]: DEBUG oslo_concurrency.lockutils [req-c0a8ee18-a1a2-4213-b5f0-c62323d38e24 req-35644b6e-5f5a-4080-be01-cefa2a813dc4 service nova] Acquired lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.688441] env[62753]: DEBUG nova.network.neutron [req-c0a8ee18-a1a2-4213-b5f0-c62323d38e24 req-35644b6e-5f5a-4080-be01-cefa2a813dc4 service nova] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Refreshing network info cache for port 25e4c4c8-9d78-4a88-9a5e-4a364ee574a5 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1077.877027] env[62753]: DEBUG nova.scheduler.client.report [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1077.898988] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332782, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1077.997276] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332785, 'name': CreateVM_Task, 'duration_secs': 0.353112} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1077.997441] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1077.998152] env[62753]: DEBUG oslo_concurrency.lockutils [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1077.998328] env[62753]: DEBUG oslo_concurrency.lockutils [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1077.998653] env[62753]: DEBUG oslo_concurrency.lockutils [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1077.998916] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3601b996-944f-4646-81bf-d72191dde1e5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.003627] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1078.003627] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52b320c3-bea4-9427-1f3f-364066098b84" [ 1078.003627] env[62753]: _type = "Task" [ 1078.003627] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.011107] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52b320c3-bea4-9427-1f3f-364066098b84, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.122580] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332786, 'name': ReconfigVM_Task, 'duration_secs': 0.204873} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.122935] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Reconfigured VM instance instance-00000062 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1078.123725] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-099fa058-b22c-4f1d-90cd-8ad030e566aa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.146939] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Reconfiguring VM instance instance-00000062 to attach disk [datastore1] eb86a978-27fa-41b2-a7e5-4b3a8ba0a152/eb86a978-27fa-41b2-a7e5-4b3a8ba0a152.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1078.147250] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-8b4e48d2-8a89-4279-ba0d-32850aa495fe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.165812] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1078.165812] env[62753]: value = "task-1332787" [ 1078.165812] env[62753]: _type = "Task" [ 1078.165812] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.176553] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332787, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.382189] env[62753]: DEBUG oslo_concurrency.lockutils [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.224s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1078.382776] env[62753]: DEBUG nova.compute.manager [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1078.385569] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 1.847s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1078.387144] env[62753]: INFO nova.compute.claims [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1078.394626] env[62753]: DEBUG nova.network.neutron [req-c0a8ee18-a1a2-4213-b5f0-c62323d38e24 req-35644b6e-5f5a-4080-be01-cefa2a813dc4 service nova] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updated VIF entry in instance network info cache for port 25e4c4c8-9d78-4a88-9a5e-4a364ee574a5. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1078.395092] env[62753]: DEBUG nova.network.neutron [req-c0a8ee18-a1a2-4213-b5f0-c62323d38e24 req-35644b6e-5f5a-4080-be01-cefa2a813dc4 service nova] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance_info_cache with network_info: [{"id": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "address": "fa:16:3e:88:6a:1d", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25e4c4c8-9d", "ovs_interfaceid": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1078.404881] env[62753]: DEBUG nova.objects.instance [None req-eaabb338-afb1-4229-ab26-2bf01b5573cf tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lazy-loading 'flavor' on Instance uuid 6e417227-895d-4576-b025-7d4a0aafa379 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1078.405607] env[62753]: DEBUG oslo_vmware.api [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332782, 'name': PowerOnVM_Task, 'duration_secs': 0.530202} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.406123] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1078.406287] env[62753]: INFO nova.compute.manager [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Took 11.18 seconds to spawn the instance on the hypervisor. [ 1078.406476] env[62753]: DEBUG nova.compute.manager [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1078.407376] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd973dc3-4140-467c-9c7f-c7cd197be312 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.513773] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52b320c3-bea4-9427-1f3f-364066098b84, 'name': SearchDatastore_Task, 'duration_secs': 0.008564} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.514082] env[62753]: DEBUG oslo_concurrency.lockutils [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.514310] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1078.514552] env[62753]: DEBUG oslo_concurrency.lockutils [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.514704] env[62753]: DEBUG oslo_concurrency.lockutils [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.514922] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1078.515198] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c50a78a5-6e9b-4bd5-b5ef-9fba743ac3c9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.524737] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1078.524936] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1078.525768] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4161af93-700c-4e99-8073-5a1f043eb8a1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1078.530782] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1078.530782] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ac0459-bac9-11eb-51f8-4040ec9b7073" [ 1078.530782] env[62753]: _type = "Task" [ 1078.530782] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1078.538435] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ac0459-bac9-11eb-51f8-4040ec9b7073, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1078.675373] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332787, 'name': ReconfigVM_Task, 'duration_secs': 0.297199} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1078.675659] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Reconfigured VM instance instance-00000062 to attach disk [datastore1] eb86a978-27fa-41b2-a7e5-4b3a8ba0a152/eb86a978-27fa-41b2-a7e5-4b3a8ba0a152.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1078.675990] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Updating instance 'eb86a978-27fa-41b2-a7e5-4b3a8ba0a152' progress to 50 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1078.709700] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1078.709854] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Starting heal instance info cache {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1078.891974] env[62753]: DEBUG nova.compute.utils [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1078.893294] env[62753]: DEBUG nova.compute.manager [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1078.893487] env[62753]: DEBUG nova.network.neutron [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1078.898974] env[62753]: DEBUG oslo_concurrency.lockutils [req-c0a8ee18-a1a2-4213-b5f0-c62323d38e24 req-35644b6e-5f5a-4080-be01-cefa2a813dc4 service nova] Releasing lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1078.909024] env[62753]: DEBUG oslo_concurrency.lockutils [None req-eaabb338-afb1-4229-ab26-2bf01b5573cf tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "refresh_cache-6e417227-895d-4576-b025-7d4a0aafa379" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1078.909024] env[62753]: DEBUG oslo_concurrency.lockutils [None req-eaabb338-afb1-4229-ab26-2bf01b5573cf tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquired lock "refresh_cache-6e417227-895d-4576-b025-7d4a0aafa379" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1078.909024] env[62753]: DEBUG nova.network.neutron [None req-eaabb338-afb1-4229-ab26-2bf01b5573cf tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1078.909024] env[62753]: DEBUG nova.objects.instance [None req-eaabb338-afb1-4229-ab26-2bf01b5573cf tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lazy-loading 'info_cache' on Instance uuid 6e417227-895d-4576-b025-7d4a0aafa379 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1078.925824] env[62753]: INFO nova.compute.manager [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Took 21.71 seconds to build instance. [ 1078.938265] env[62753]: DEBUG nova.policy [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2a7f23e0237349bdbebc6135191ffe9c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2dd81bdb01ff4c39a4959f4af2acf61a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1079.040655] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52ac0459-bac9-11eb-51f8-4040ec9b7073, 'name': SearchDatastore_Task, 'duration_secs': 0.007415} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.041445] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-38cd9c28-d057-407e-9d47-cabbd78a842a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.046820] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1079.046820] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d58cf2-c75a-332a-438c-c9eaa8e8664f" [ 1079.046820] env[62753]: _type = "Task" [ 1079.046820] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.057099] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d58cf2-c75a-332a-438c-c9eaa8e8664f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.180470] env[62753]: DEBUG nova.network.neutron [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Successfully created port: dd52d78f-614d-4686-a0aa-f92598ea6932 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1079.184449] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-986baf2a-5e3f-4df3-b2cf-a4579e78e0d2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.206558] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c97c175-4b3d-4477-a879-9a93e4d00ecc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.225139] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Updating instance 'eb86a978-27fa-41b2-a7e5-4b3a8ba0a152' progress to 67 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1079.396534] env[62753]: DEBUG nova.compute.manager [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1079.411374] env[62753]: DEBUG nova.objects.base [None req-eaabb338-afb1-4229-ab26-2bf01b5573cf tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Object Instance<6e417227-895d-4576-b025-7d4a0aafa379> lazy-loaded attributes: flavor,info_cache {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1079.428526] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6727a771-adfc-4dca-8964-91e60a28bfc4 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.221s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1079.548362] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-808ac933-bf86-4f65-b574-2a95ef099052 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.559286] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52d58cf2-c75a-332a-438c-c9eaa8e8664f, 'name': SearchDatastore_Task, 'duration_secs': 0.010173} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1079.560993] env[62753]: DEBUG oslo_concurrency.lockutils [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1079.561742] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 22cea165-cfde-403f-a0e2-82a86f5baa51/22cea165-cfde-403f-a0e2-82a86f5baa51.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1079.561742] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-489f51ab-aa53-43ad-97ea-38010099f3f3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.563921] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1d38ad99-57e0-4f75-8773-5a3324f9ccad {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.596371] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fdf7ed64-fbb4-4850-a5da-98c10d75d4d9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.598822] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1079.598822] env[62753]: value = "task-1332788" [ 1079.598822] env[62753]: _type = "Task" [ 1079.598822] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1079.605208] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f85e0ae7-4523-4d71-b463-26a0bdb88b09 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1079.611352] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332788, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1079.620662] env[62753]: DEBUG nova.compute.provider_tree [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1079.772445] env[62753]: DEBUG nova.network.neutron [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Port 93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9 binding to destination host cpu-1 is already ACTIVE {{(pid=62753) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1080.005803] env[62753]: DEBUG nova.compute.manager [req-02116b6e-d9df-4215-ba4c-61529c131fd8 req-ef9fb037-ddd3-4ca2-b229-c61b352bf828 service nova] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Received event network-changed-ad95cf04-2066-4452-b160-65cf3cddbe87 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1080.006205] env[62753]: DEBUG nova.compute.manager [req-02116b6e-d9df-4215-ba4c-61529c131fd8 req-ef9fb037-ddd3-4ca2-b229-c61b352bf828 service nova] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Refreshing instance network info cache due to event network-changed-ad95cf04-2066-4452-b160-65cf3cddbe87. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1080.006562] env[62753]: DEBUG oslo_concurrency.lockutils [req-02116b6e-d9df-4215-ba4c-61529c131fd8 req-ef9fb037-ddd3-4ca2-b229-c61b352bf828 service nova] Acquiring lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1080.006763] env[62753]: DEBUG oslo_concurrency.lockutils [req-02116b6e-d9df-4215-ba4c-61529c131fd8 req-ef9fb037-ddd3-4ca2-b229-c61b352bf828 service nova] Acquired lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1080.007011] env[62753]: DEBUG nova.network.neutron [req-02116b6e-d9df-4215-ba4c-61529c131fd8 req-ef9fb037-ddd3-4ca2-b229-c61b352bf828 service nova] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Refreshing network info cache for port ad95cf04-2066-4452-b160-65cf3cddbe87 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1080.110035] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332788, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.123857] env[62753]: DEBUG nova.scheduler.client.report [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1080.155878] env[62753]: DEBUG nova.network.neutron [None req-eaabb338-afb1-4229-ab26-2bf01b5573cf tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Updating instance_info_cache with network_info: [{"id": "2254df7c-8da6-4c8d-ae9c-5b10e79dc88f", "address": "fa:16:3e:05:70:9f", "network": {"id": "e3de96b6-0e52-4885-8747-d3e60da74ef2", "bridge": "br-int", "label": "tempest-ListServerFiltersTestJSON-91543986-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "7b76b178732e41688f767fdd65934193", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "13d625c9-77ec-4edb-a56b-9f37a314cc39", "external-id": "nsx-vlan-transportzone-358", "segmentation_id": 358, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap2254df7c-8d", "ovs_interfaceid": "2254df7c-8da6-4c8d-ae9c-5b10e79dc88f", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1080.410955] env[62753]: DEBUG nova.compute.manager [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1080.438304] env[62753]: DEBUG nova.virt.hardware [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1080.438304] env[62753]: DEBUG nova.virt.hardware [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1080.438304] env[62753]: DEBUG nova.virt.hardware [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1080.438583] env[62753]: DEBUG nova.virt.hardware [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1080.438583] env[62753]: DEBUG nova.virt.hardware [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1080.438718] env[62753]: DEBUG nova.virt.hardware [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1080.438983] env[62753]: DEBUG nova.virt.hardware [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1080.439247] env[62753]: DEBUG nova.virt.hardware [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1080.439442] env[62753]: DEBUG nova.virt.hardware [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1080.439619] env[62753]: DEBUG nova.virt.hardware [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1080.439802] env[62753]: DEBUG nova.virt.hardware [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1080.440705] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cec70343-f7c4-4f60-a5f6-c360398d36c6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.448756] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b8392d77-4315-41ac-b0ce-fc63d8ba36a7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.612574] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332788, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.557716} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1080.612574] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 22cea165-cfde-403f-a0e2-82a86f5baa51/22cea165-cfde-403f-a0e2-82a86f5baa51.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1080.612574] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1080.612782] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b29e0a18-4629-46a0-ba0b-a509fe870dbd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1080.617528] env[62753]: DEBUG nova.compute.manager [req-bec2b6d1-675a-4984-a758-f366d6164adf req-23085b2b-29e9-47d2-be33-9202e3447bc3 service nova] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Received event network-vif-plugged-dd52d78f-614d-4686-a0aa-f92598ea6932 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1080.617724] env[62753]: DEBUG oslo_concurrency.lockutils [req-bec2b6d1-675a-4984-a758-f366d6164adf req-23085b2b-29e9-47d2-be33-9202e3447bc3 service nova] Acquiring lock "a2e7fa98-566e-40f1-8da7-9318de89c14c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.617901] env[62753]: DEBUG oslo_concurrency.lockutils [req-bec2b6d1-675a-4984-a758-f366d6164adf req-23085b2b-29e9-47d2-be33-9202e3447bc3 service nova] Lock "a2e7fa98-566e-40f1-8da7-9318de89c14c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.618085] env[62753]: DEBUG oslo_concurrency.lockutils [req-bec2b6d1-675a-4984-a758-f366d6164adf req-23085b2b-29e9-47d2-be33-9202e3447bc3 service nova] Lock "a2e7fa98-566e-40f1-8da7-9318de89c14c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.618254] env[62753]: DEBUG nova.compute.manager [req-bec2b6d1-675a-4984-a758-f366d6164adf req-23085b2b-29e9-47d2-be33-9202e3447bc3 service nova] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] No waiting events found dispatching network-vif-plugged-dd52d78f-614d-4686-a0aa-f92598ea6932 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1080.618420] env[62753]: WARNING nova.compute.manager [req-bec2b6d1-675a-4984-a758-f366d6164adf req-23085b2b-29e9-47d2-be33-9202e3447bc3 service nova] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Received unexpected event network-vif-plugged-dd52d78f-614d-4686-a0aa-f92598ea6932 for instance with vm_state building and task_state spawning. [ 1080.624218] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1080.624218] env[62753]: value = "task-1332789" [ 1080.624218] env[62753]: _type = "Task" [ 1080.624218] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1080.630400] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.245s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.630891] env[62753]: DEBUG nova.compute.manager [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1080.639186] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332789, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1080.658714] env[62753]: DEBUG oslo_concurrency.lockutils [None req-eaabb338-afb1-4229-ab26-2bf01b5573cf tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Releasing lock "refresh_cache-6e417227-895d-4576-b025-7d4a0aafa379" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1080.792228] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1080.792574] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1080.793229] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1080.804039] env[62753]: DEBUG nova.network.neutron [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Successfully updated port: dd52d78f-614d-4686-a0aa-f92598ea6932 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1080.854981] env[62753]: DEBUG nova.network.neutron [req-02116b6e-d9df-4215-ba4c-61529c131fd8 req-ef9fb037-ddd3-4ca2-b229-c61b352bf828 service nova] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updated VIF entry in instance network info cache for port ad95cf04-2066-4452-b160-65cf3cddbe87. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1080.855389] env[62753]: DEBUG nova.network.neutron [req-02116b6e-d9df-4215-ba4c-61529c131fd8 req-ef9fb037-ddd3-4ca2-b229-c61b352bf828 service nova] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance_info_cache with network_info: [{"id": "ad95cf04-2066-4452-b160-65cf3cddbe87", "address": "fa:16:3e:eb:5e:01", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad95cf04-20", "ovs_interfaceid": "ad95cf04-2066-4452-b160-65cf3cddbe87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1081.135744] env[62753]: DEBUG nova.compute.utils [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1081.137155] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332789, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.071155} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1081.137630] env[62753]: DEBUG nova.compute.manager [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1081.137801] env[62753]: DEBUG nova.network.neutron [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1081.139542] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1081.140385] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4664edf-6a79-46b4-a61e-243e7b6707af {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.163184] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 22cea165-cfde-403f-a0e2-82a86f5baa51/22cea165-cfde-403f-a0e2-82a86f5baa51.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1081.164085] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c425828b-d7eb-417b-951e-4b52d1438fe5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.178975] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaabb338-afb1-4229-ab26-2bf01b5573cf tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1081.179584] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-23c990db-b568-4b30-8e48-d1c3db0be2f5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1081.187501] env[62753]: DEBUG nova.policy [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2a7f23e0237349bdbebc6135191ffe9c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2dd81bdb01ff4c39a4959f4af2acf61a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1081.191790] env[62753]: DEBUG oslo_vmware.api [None req-eaabb338-afb1-4229-ab26-2bf01b5573cf tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1081.191790] env[62753]: value = "task-1332791" [ 1081.191790] env[62753]: _type = "Task" [ 1081.191790] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.192042] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1081.192042] env[62753]: value = "task-1332790" [ 1081.192042] env[62753]: _type = "Task" [ 1081.192042] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1081.203427] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332790, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.206481] env[62753]: DEBUG oslo_vmware.api [None req-eaabb338-afb1-4229-ab26-2bf01b5573cf tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332791, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.305976] env[62753]: DEBUG oslo_concurrency.lockutils [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "refresh_cache-a2e7fa98-566e-40f1-8da7-9318de89c14c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.306356] env[62753]: DEBUG oslo_concurrency.lockutils [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired lock "refresh_cache-a2e7fa98-566e-40f1-8da7-9318de89c14c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.306356] env[62753]: DEBUG nova.network.neutron [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1081.358690] env[62753]: DEBUG oslo_concurrency.lockutils [req-02116b6e-d9df-4215-ba4c-61529c131fd8 req-ef9fb037-ddd3-4ca2-b229-c61b352bf828 service nova] Releasing lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1081.441386] env[62753]: DEBUG nova.network.neutron [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Successfully created port: 450be489-a9db-47f3-91e7-c6673a8af916 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1081.640911] env[62753]: DEBUG nova.compute.manager [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1081.705618] env[62753]: DEBUG oslo_vmware.api [None req-eaabb338-afb1-4229-ab26-2bf01b5573cf tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332791, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.708921] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332790, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1081.854282] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "refresh_cache-eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1081.854496] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "refresh_cache-eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1081.854681] env[62753]: DEBUG nova.network.neutron [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1081.857515] env[62753]: DEBUG nova.network.neutron [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1081.982548] env[62753]: DEBUG nova.network.neutron [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Updating instance_info_cache with network_info: [{"id": "dd52d78f-614d-4686-a0aa-f92598ea6932", "address": "fa:16:3e:a7:28:24", "network": {"id": "939bb8aa-f8d4-44cb-960f-c5a5fbcf99c8", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1149502111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dd81bdb01ff4c39a4959f4af2acf61a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd52d78f-61", "ovs_interfaceid": "dd52d78f-614d-4686-a0aa-f92598ea6932", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.205216] env[62753]: DEBUG oslo_vmware.api [None req-eaabb338-afb1-4229-ab26-2bf01b5573cf tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332791, 'name': PowerOnVM_Task, 'duration_secs': 0.647395} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.208438] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-eaabb338-afb1-4229-ab26-2bf01b5573cf tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1082.208637] env[62753]: DEBUG nova.compute.manager [None req-eaabb338-afb1-4229-ab26-2bf01b5573cf tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1082.208967] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332790, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.209673] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-179139d4-5498-4130-a31b-c2e945134ad0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.485871] env[62753]: DEBUG oslo_concurrency.lockutils [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Releasing lock "refresh_cache-a2e7fa98-566e-40f1-8da7-9318de89c14c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1082.486235] env[62753]: DEBUG nova.compute.manager [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Instance network_info: |[{"id": "dd52d78f-614d-4686-a0aa-f92598ea6932", "address": "fa:16:3e:a7:28:24", "network": {"id": "939bb8aa-f8d4-44cb-960f-c5a5fbcf99c8", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1149502111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dd81bdb01ff4c39a4959f4af2acf61a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd52d78f-61", "ovs_interfaceid": "dd52d78f-614d-4686-a0aa-f92598ea6932", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1082.486852] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:a7:28:24', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'dd52d78f-614d-4686-a0aa-f92598ea6932', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1082.494706] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Creating folder: Project (2dd81bdb01ff4c39a4959f4af2acf61a). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1082.495085] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-84414cc1-04da-41c7-b777-6dab14cc8eaf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.505835] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Created folder: Project (2dd81bdb01ff4c39a4959f4af2acf61a) in parent group-v284541. [ 1082.506102] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Creating folder: Instances. Parent ref: group-v284685. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1082.506357] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-505341b9-53c3-4754-b6da-7bda4ba08e48 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.515310] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Created folder: Instances in parent group-v284685. [ 1082.515543] env[62753]: DEBUG oslo.service.loopingcall [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1082.515736] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1082.515953] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-5b6c432e-3c54-4cf4-a965-e1a1766157f7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.535528] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1082.535528] env[62753]: value = "task-1332794" [ 1082.535528] env[62753]: _type = "Task" [ 1082.535528] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.542733] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332794, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.544628] env[62753]: DEBUG nova.network.neutron [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Updating instance_info_cache with network_info: [{"id": "93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9", "address": "fa:16:3e:90:68:e2", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93d26d06-fe", "ovs_interfaceid": "93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1082.650812] env[62753]: DEBUG nova.compute.manager [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1082.682365] env[62753]: DEBUG nova.virt.hardware [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1082.682650] env[62753]: DEBUG nova.virt.hardware [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1082.682826] env[62753]: DEBUG nova.virt.hardware [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1082.683017] env[62753]: DEBUG nova.virt.hardware [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1082.683176] env[62753]: DEBUG nova.virt.hardware [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1082.683327] env[62753]: DEBUG nova.virt.hardware [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1082.683538] env[62753]: DEBUG nova.virt.hardware [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1082.683699] env[62753]: DEBUG nova.virt.hardware [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1082.683872] env[62753]: DEBUG nova.virt.hardware [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1082.684052] env[62753]: DEBUG nova.virt.hardware [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1082.684238] env[62753]: DEBUG nova.virt.hardware [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1082.685203] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8962acb2-bb99-4caa-bd44-dd8ab03c626d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.693913] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7471463c-e4d7-4127-8ec7-870d6e0aa72a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.706886] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332790, 'name': ReconfigVM_Task, 'duration_secs': 1.474937} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1082.714788] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 22cea165-cfde-403f-a0e2-82a86f5baa51/22cea165-cfde-403f-a0e2-82a86f5baa51.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1082.715504] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-26f08513-f83e-4043-b5a1-a12f1000bc9d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1082.720860] env[62753]: DEBUG nova.compute.manager [req-a50d8e01-4305-4b18-a5a7-7af6b12513b4 req-620827f0-e4f0-4041-8f67-5b731c6c4d23 service nova] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Received event network-changed-dd52d78f-614d-4686-a0aa-f92598ea6932 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1082.721067] env[62753]: DEBUG nova.compute.manager [req-a50d8e01-4305-4b18-a5a7-7af6b12513b4 req-620827f0-e4f0-4041-8f67-5b731c6c4d23 service nova] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Refreshing instance network info cache due to event network-changed-dd52d78f-614d-4686-a0aa-f92598ea6932. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1082.721286] env[62753]: DEBUG oslo_concurrency.lockutils [req-a50d8e01-4305-4b18-a5a7-7af6b12513b4 req-620827f0-e4f0-4041-8f67-5b731c6c4d23 service nova] Acquiring lock "refresh_cache-a2e7fa98-566e-40f1-8da7-9318de89c14c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1082.721434] env[62753]: DEBUG oslo_concurrency.lockutils [req-a50d8e01-4305-4b18-a5a7-7af6b12513b4 req-620827f0-e4f0-4041-8f67-5b731c6c4d23 service nova] Acquired lock "refresh_cache-a2e7fa98-566e-40f1-8da7-9318de89c14c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1082.721727] env[62753]: DEBUG nova.network.neutron [req-a50d8e01-4305-4b18-a5a7-7af6b12513b4 req-620827f0-e4f0-4041-8f67-5b731c6c4d23 service nova] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Refreshing network info cache for port dd52d78f-614d-4686-a0aa-f92598ea6932 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1082.732205] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1082.732205] env[62753]: value = "task-1332795" [ 1082.732205] env[62753]: _type = "Task" [ 1082.732205] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1082.743316] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332795, 'name': Rename_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1082.749778] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Didn't find any instances for network info cache update. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1082.993352] env[62753]: DEBUG nova.network.neutron [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Successfully updated port: 450be489-a9db-47f3-91e7-c6673a8af916 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1083.047091] env[62753]: DEBUG oslo_concurrency.lockutils [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "refresh_cache-eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.050735] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332794, 'name': CreateVM_Task, 'duration_secs': 0.363487} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.051103] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1083.051785] env[62753]: DEBUG oslo_concurrency.lockutils [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.052012] env[62753]: DEBUG oslo_concurrency.lockutils [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.052458] env[62753]: DEBUG oslo_concurrency.lockutils [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1083.052800] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a3cd565f-e58e-4ae7-8c97-6270bb977db7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.057816] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1083.057816] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52479378-8ef9-aa33-e2ea-14079414db7a" [ 1083.057816] env[62753]: _type = "Task" [ 1083.057816] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.065541] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52479378-8ef9-aa33-e2ea-14079414db7a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.241905] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332795, 'name': Rename_Task, 'duration_secs': 0.142131} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.242381] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1083.243085] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-423ea81d-1cd1-4acf-9a46-4af266e0be5a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.249904] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1083.249904] env[62753]: value = "task-1332796" [ 1083.249904] env[62753]: _type = "Task" [ 1083.249904] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.259175] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332796, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.422389] env[62753]: DEBUG nova.network.neutron [req-a50d8e01-4305-4b18-a5a7-7af6b12513b4 req-620827f0-e4f0-4041-8f67-5b731c6c4d23 service nova] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Updated VIF entry in instance network info cache for port dd52d78f-614d-4686-a0aa-f92598ea6932. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1083.422956] env[62753]: DEBUG nova.network.neutron [req-a50d8e01-4305-4b18-a5a7-7af6b12513b4 req-620827f0-e4f0-4041-8f67-5b731c6c4d23 service nova] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Updating instance_info_cache with network_info: [{"id": "dd52d78f-614d-4686-a0aa-f92598ea6932", "address": "fa:16:3e:a7:28:24", "network": {"id": "939bb8aa-f8d4-44cb-960f-c5a5fbcf99c8", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1149502111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dd81bdb01ff4c39a4959f4af2acf61a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd52d78f-61", "ovs_interfaceid": "dd52d78f-614d-4686-a0aa-f92598ea6932", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1083.497341] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "refresh_cache-51790f51-c464-48cc-9891-41cae5a8f63e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.497341] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired lock "refresh_cache-51790f51-c464-48cc-9891-41cae5a8f63e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.497341] env[62753]: DEBUG nova.network.neutron [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1083.570095] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52479378-8ef9-aa33-e2ea-14079414db7a, 'name': SearchDatastore_Task, 'duration_secs': 0.009711} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.571262] env[62753]: DEBUG oslo_concurrency.lockutils [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1083.571520] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1083.571771] env[62753]: DEBUG oslo_concurrency.lockutils [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1083.571937] env[62753]: DEBUG oslo_concurrency.lockutils [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1083.572154] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1083.573035] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9886ece-ba56-415e-b4cd-9694dad2a034 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.576293] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c74e05e1-44b7-4a33-80dc-650c3a8418b5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.596385] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-62652800-84e1-4566-a71a-f96c280d391b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.600248] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1083.600248] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1083.600390] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a03ea13e-8a05-4222-9256-e1115c141479 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.608501] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1083.608501] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5202fee0-89a9-ae72-10c8-9b8b6485f3a8" [ 1083.608501] env[62753]: _type = "Task" [ 1083.608501] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.608854] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Updating instance 'eb86a978-27fa-41b2-a7e5-4b3a8ba0a152' progress to 83 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1083.621126] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5202fee0-89a9-ae72-10c8-9b8b6485f3a8, 'name': SearchDatastore_Task, 'duration_secs': 0.008858} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1083.621954] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d6107d5e-9126-47f2-9ed1-47dd0f20f436 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1083.627383] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1083.627383] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52183957-1857-a6ce-9a4c-22bcb4f76615" [ 1083.627383] env[62753]: _type = "Task" [ 1083.627383] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1083.635849] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52183957-1857-a6ce-9a4c-22bcb4f76615, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.759869] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332796, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1083.926288] env[62753]: DEBUG oslo_concurrency.lockutils [req-a50d8e01-4305-4b18-a5a7-7af6b12513b4 req-620827f0-e4f0-4041-8f67-5b731c6c4d23 service nova] Releasing lock "refresh_cache-a2e7fa98-566e-40f1-8da7-9318de89c14c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.031660] env[62753]: DEBUG nova.network.neutron [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1084.118375] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1084.118687] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4e4f222f-432c-419a-9332-16dc31b16aa0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.125146] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1084.125146] env[62753]: value = "task-1332797" [ 1084.125146] env[62753]: _type = "Task" [ 1084.125146] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.136765] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332797, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.141174] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52183957-1857-a6ce-9a4c-22bcb4f76615, 'name': SearchDatastore_Task, 'duration_secs': 0.009355} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1084.141432] env[62753]: DEBUG oslo_concurrency.lockutils [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.141717] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] a2e7fa98-566e-40f1-8da7-9318de89c14c/a2e7fa98-566e-40f1-8da7-9318de89c14c.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1084.141992] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-bc6a3db5-3d03-452d-8f55-36e7b9f42cf5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.147923] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1084.147923] env[62753]: value = "task-1332798" [ 1084.147923] env[62753]: _type = "Task" [ 1084.147923] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.155400] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332798, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.173885] env[62753]: DEBUG nova.network.neutron [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Updating instance_info_cache with network_info: [{"id": "450be489-a9db-47f3-91e7-c6673a8af916", "address": "fa:16:3e:f1:67:9b", "network": {"id": "939bb8aa-f8d4-44cb-960f-c5a5fbcf99c8", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1149502111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dd81bdb01ff4c39a4959f4af2acf61a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap450be489-a9", "ovs_interfaceid": "450be489-a9db-47f3-91e7-c6673a8af916", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1084.262527] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332796, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.636480] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332797, 'name': PowerOnVM_Task} progress is 81%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.658438] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332798, 'name': CopyVirtualDisk_Task} progress is 77%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.676261] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Releasing lock "refresh_cache-51790f51-c464-48cc-9891-41cae5a8f63e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1084.676642] env[62753]: DEBUG nova.compute.manager [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Instance network_info: |[{"id": "450be489-a9db-47f3-91e7-c6673a8af916", "address": "fa:16:3e:f1:67:9b", "network": {"id": "939bb8aa-f8d4-44cb-960f-c5a5fbcf99c8", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1149502111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dd81bdb01ff4c39a4959f4af2acf61a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap450be489-a9", "ovs_interfaceid": "450be489-a9db-47f3-91e7-c6673a8af916", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1084.677067] env[62753]: DEBUG oslo_concurrency.lockutils [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.677344] env[62753]: DEBUG oslo_concurrency.lockutils [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.677565] env[62753]: DEBUG oslo_concurrency.lockutils [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.677813] env[62753]: DEBUG oslo_concurrency.lockutils [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.678010] env[62753]: DEBUG oslo_concurrency.lockutils [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.679881] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:f1:67:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '450be489-a9db-47f3-91e7-c6673a8af916', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1084.687793] env[62753]: DEBUG oslo.service.loopingcall [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1084.688357] env[62753]: INFO nova.compute.manager [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Terminating instance [ 1084.689769] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1084.690370] env[62753]: DEBUG nova.compute.manager [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1084.690821] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1084.691301] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-6f852b93-94b7-4653-8ea3-a2a1be35572d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.706264] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f64131a7-c837-4dcf-92f2-25a210d5be0a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.713819] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1084.714952] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-19b3feea-82b6-469d-b0a2-837fc7479423 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1084.716232] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1084.716232] env[62753]: value = "task-1332799" [ 1084.716232] env[62753]: _type = "Task" [ 1084.716232] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.720381] env[62753]: DEBUG oslo_vmware.api [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1084.720381] env[62753]: value = "task-1332800" [ 1084.720381] env[62753]: _type = "Task" [ 1084.720381] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1084.725948] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332799, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.730645] env[62753]: DEBUG oslo_vmware.api [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332800, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1084.747455] env[62753]: DEBUG nova.compute.manager [req-a429cdd0-f9c0-4992-9239-07694c2c9336 req-a243865f-8da9-4ecd-ba60-5d65aa2616fb service nova] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Received event network-vif-plugged-450be489-a9db-47f3-91e7-c6673a8af916 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1084.747638] env[62753]: DEBUG oslo_concurrency.lockutils [req-a429cdd0-f9c0-4992-9239-07694c2c9336 req-a243865f-8da9-4ecd-ba60-5d65aa2616fb service nova] Acquiring lock "51790f51-c464-48cc-9891-41cae5a8f63e-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1084.747859] env[62753]: DEBUG oslo_concurrency.lockutils [req-a429cdd0-f9c0-4992-9239-07694c2c9336 req-a243865f-8da9-4ecd-ba60-5d65aa2616fb service nova] Lock "51790f51-c464-48cc-9891-41cae5a8f63e-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1084.748046] env[62753]: DEBUG oslo_concurrency.lockutils [req-a429cdd0-f9c0-4992-9239-07694c2c9336 req-a243865f-8da9-4ecd-ba60-5d65aa2616fb service nova] Lock "51790f51-c464-48cc-9891-41cae5a8f63e-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1084.748250] env[62753]: DEBUG nova.compute.manager [req-a429cdd0-f9c0-4992-9239-07694c2c9336 req-a243865f-8da9-4ecd-ba60-5d65aa2616fb service nova] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] No waiting events found dispatching network-vif-plugged-450be489-a9db-47f3-91e7-c6673a8af916 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1084.748449] env[62753]: WARNING nova.compute.manager [req-a429cdd0-f9c0-4992-9239-07694c2c9336 req-a243865f-8da9-4ecd-ba60-5d65aa2616fb service nova] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Received unexpected event network-vif-plugged-450be489-a9db-47f3-91e7-c6673a8af916 for instance with vm_state building and task_state spawning. [ 1084.748624] env[62753]: DEBUG nova.compute.manager [req-a429cdd0-f9c0-4992-9239-07694c2c9336 req-a243865f-8da9-4ecd-ba60-5d65aa2616fb service nova] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Received event network-changed-450be489-a9db-47f3-91e7-c6673a8af916 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1084.748793] env[62753]: DEBUG nova.compute.manager [req-a429cdd0-f9c0-4992-9239-07694c2c9336 req-a243865f-8da9-4ecd-ba60-5d65aa2616fb service nova] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Refreshing instance network info cache due to event network-changed-450be489-a9db-47f3-91e7-c6673a8af916. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1084.749014] env[62753]: DEBUG oslo_concurrency.lockutils [req-a429cdd0-f9c0-4992-9239-07694c2c9336 req-a243865f-8da9-4ecd-ba60-5d65aa2616fb service nova] Acquiring lock "refresh_cache-51790f51-c464-48cc-9891-41cae5a8f63e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1084.749168] env[62753]: DEBUG oslo_concurrency.lockutils [req-a429cdd0-f9c0-4992-9239-07694c2c9336 req-a243865f-8da9-4ecd-ba60-5d65aa2616fb service nova] Acquired lock "refresh_cache-51790f51-c464-48cc-9891-41cae5a8f63e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1084.749329] env[62753]: DEBUG nova.network.neutron [req-a429cdd0-f9c0-4992-9239-07694c2c9336 req-a243865f-8da9-4ecd-ba60-5d65aa2616fb service nova] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Refreshing network info cache for port 450be489-a9db-47f3-91e7-c6673a8af916 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1084.762123] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332796, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.137372] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332797, 'name': PowerOnVM_Task} progress is 81%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.157495] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332798, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.598466} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.160417] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] a2e7fa98-566e-40f1-8da7-9318de89c14c/a2e7fa98-566e-40f1-8da7-9318de89c14c.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1085.160417] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1085.160417] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-f0c6b1c4-ab99-4e23-9974-1f84b06f3c81 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.165039] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1085.165039] env[62753]: value = "task-1332801" [ 1085.165039] env[62753]: _type = "Task" [ 1085.165039] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.172072] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332801, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.227357] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332799, 'name': CreateVM_Task, 'duration_secs': 0.403521} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.227969] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1085.228659] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.228846] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.229258] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1085.232129] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5951148a-b4d5-4f66-9b24-935aecbb0b69 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.233570] env[62753]: DEBUG oslo_vmware.api [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332800, 'name': PowerOffVM_Task, 'duration_secs': 0.231658} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.233815] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1085.233988] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1085.234509] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a7fea1ef-8358-4fdc-8540-3fd3c188d47f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.236852] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1085.236852] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52883438-3b72-eabe-35e3-159e6284d1be" [ 1085.236852] env[62753]: _type = "Task" [ 1085.236852] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.244067] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52883438-3b72-eabe-35e3-159e6284d1be, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.260253] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332796, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.299097] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1085.299335] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1085.299524] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Deleting the datastore file [datastore2] 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1085.299819] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-d1888e5c-6bea-463f-8aaf-8d2a2bf41eb6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.305690] env[62753]: DEBUG oslo_vmware.api [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1085.305690] env[62753]: value = "task-1332803" [ 1085.305690] env[62753]: _type = "Task" [ 1085.305690] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.314647] env[62753]: DEBUG oslo_vmware.api [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332803, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.452135] env[62753]: DEBUG nova.network.neutron [req-a429cdd0-f9c0-4992-9239-07694c2c9336 req-a243865f-8da9-4ecd-ba60-5d65aa2616fb service nova] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Updated VIF entry in instance network info cache for port 450be489-a9db-47f3-91e7-c6673a8af916. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1085.452523] env[62753]: DEBUG nova.network.neutron [req-a429cdd0-f9c0-4992-9239-07694c2c9336 req-a243865f-8da9-4ecd-ba60-5d65aa2616fb service nova] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Updating instance_info_cache with network_info: [{"id": "450be489-a9db-47f3-91e7-c6673a8af916", "address": "fa:16:3e:f1:67:9b", "network": {"id": "939bb8aa-f8d4-44cb-960f-c5a5fbcf99c8", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1149502111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dd81bdb01ff4c39a4959f4af2acf61a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap450be489-a9", "ovs_interfaceid": "450be489-a9db-47f3-91e7-c6673a8af916", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1085.636668] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332797, 'name': PowerOnVM_Task} progress is 81%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.674735] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332801, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064133} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.675078] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1085.675822] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63a5151a-6691-4ed6-90b2-d55f074dd204 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.697244] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] a2e7fa98-566e-40f1-8da7-9318de89c14c/a2e7fa98-566e-40f1-8da7-9318de89c14c.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1085.697499] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c50bfb26-76e0-40a4-aa48-3a8b1f38dacf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.717672] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1085.717672] env[62753]: value = "task-1332804" [ 1085.717672] env[62753]: _type = "Task" [ 1085.717672] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.725757] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332804, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.746668] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52883438-3b72-eabe-35e3-159e6284d1be, 'name': SearchDatastore_Task, 'duration_secs': 0.055435} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.746985] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1085.747265] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1085.747520] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1085.747670] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1085.747849] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1085.748131] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-d4bee38c-1c60-4ae6-ac91-05252a9a76a4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.757323] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1085.757504] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1085.758490] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6bcfec3c-67af-4353-9d37-0813667891bb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1085.763617] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332796, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.766446] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1085.766446] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52611952-8789-1a47-a888-cbeb03493423" [ 1085.766446] env[62753]: _type = "Task" [ 1085.766446] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1085.773485] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52611952-8789-1a47-a888-cbeb03493423, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1085.814942] env[62753]: DEBUG oslo_vmware.api [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332803, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.486008} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1085.815270] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1085.815511] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1085.815728] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1085.815913] env[62753]: INFO nova.compute.manager [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1085.816220] env[62753]: DEBUG oslo.service.loopingcall [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1085.816426] env[62753]: DEBUG nova.compute.manager [-] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1085.816524] env[62753]: DEBUG nova.network.neutron [-] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1085.956020] env[62753]: DEBUG oslo_concurrency.lockutils [req-a429cdd0-f9c0-4992-9239-07694c2c9336 req-a243865f-8da9-4ecd-ba60-5d65aa2616fb service nova] Releasing lock "refresh_cache-51790f51-c464-48cc-9891-41cae5a8f63e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.141299] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332797, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.230133] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332804, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.267245] env[62753]: DEBUG oslo_vmware.api [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332796, 'name': PowerOnVM_Task, 'duration_secs': 2.942431} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.270752] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1086.270898] env[62753]: INFO nova.compute.manager [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Took 10.90 seconds to spawn the instance on the hypervisor. [ 1086.271163] env[62753]: DEBUG nova.compute.manager [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1086.271958] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c24def78-0838-4585-a363-9f85cf1915f9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.284976] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52611952-8789-1a47-a888-cbeb03493423, 'name': SearchDatastore_Task, 'duration_secs': 0.010159} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.286114] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ca9fdd9e-d699-47d7-b03a-40889d9686fa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.292577] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1086.292577] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5217f465-2bf9-0b6c-34a6-07d653998c1f" [ 1086.292577] env[62753]: _type = "Task" [ 1086.292577] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.303287] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5217f465-2bf9-0b6c-34a6-07d653998c1f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.576409] env[62753]: DEBUG nova.network.neutron [-] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1086.638404] env[62753]: DEBUG oslo_vmware.api [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332797, 'name': PowerOnVM_Task, 'duration_secs': 2.138879} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.638774] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1086.638849] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-aef059b4-3005-4057-b8d0-d7a5008c7b3a tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Updating instance 'eb86a978-27fa-41b2-a7e5-4b3a8ba0a152' progress to 100 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1086.728149] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332804, 'name': ReconfigVM_Task, 'duration_secs': 0.55136} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.728435] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Reconfigured VM instance instance-00000069 to attach disk [datastore2] a2e7fa98-566e-40f1-8da7-9318de89c14c/a2e7fa98-566e-40f1-8da7-9318de89c14c.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1086.729050] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b9b8d0a7-3980-40f2-8f6e-2a6ae37730bb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.735913] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1086.735913] env[62753]: value = "task-1332805" [ 1086.735913] env[62753]: _type = "Task" [ 1086.735913] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.743433] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332805, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1086.773858] env[62753]: DEBUG nova.compute.manager [req-84bde937-91f0-4c42-bd7c-9b625ccd0ec7 req-03deb311-0a7e-4c20-9e60-e67a3620ac04 service nova] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Received event network-vif-deleted-3f1e92c2-bb9a-4d5d-a2ea-8f5943d15260 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1086.792749] env[62753]: INFO nova.compute.manager [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Took 22.13 seconds to build instance. [ 1086.805095] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5217f465-2bf9-0b6c-34a6-07d653998c1f, 'name': SearchDatastore_Task, 'duration_secs': 0.015278} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1086.805950] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1086.806296] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 51790f51-c464-48cc-9891-41cae5a8f63e/51790f51-c464-48cc-9891-41cae5a8f63e.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1086.806859] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-50911b5d-542d-41b6-85b9-6aefa8733409 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1086.815275] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1086.815275] env[62753]: value = "task-1332806" [ 1086.815275] env[62753]: _type = "Task" [ 1086.815275] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1086.824581] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332806, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.079155] env[62753]: INFO nova.compute.manager [-] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Took 1.26 seconds to deallocate network for instance. [ 1087.247630] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332805, 'name': Rename_Task, 'duration_secs': 0.205842} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.247805] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1087.247975] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c20cf21a-abad-40ca-aab9-a4c899aed0c2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.256271] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1087.256271] env[62753]: value = "task-1332807" [ 1087.256271] env[62753]: _type = "Task" [ 1087.256271] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.269928] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332807, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.299301] env[62753]: DEBUG oslo_concurrency.lockutils [None req-21b693d8-4be7-4d55-9d04-b11ad41885ff tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 23.654s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1087.329793] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332806, 'name': CopyVirtualDisk_Task} progress is 51%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.585238] env[62753]: DEBUG oslo_concurrency.lockutils [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1087.585819] env[62753]: DEBUG oslo_concurrency.lockutils [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1087.585960] env[62753]: DEBUG nova.objects.instance [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lazy-loading 'resources' on Instance uuid 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1087.769114] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332807, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.829735] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332806, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.643642} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1087.830227] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 51790f51-c464-48cc-9891-41cae5a8f63e/51790f51-c464-48cc-9891-41cae5a8f63e.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1087.830433] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1087.830821] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-fdca8911-39e5-4579-ae93-6cd2adb3eb5d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1087.839553] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1087.839553] env[62753]: value = "task-1332808" [ 1087.839553] env[62753]: _type = "Task" [ 1087.839553] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1087.850260] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332808, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1087.883897] env[62753]: DEBUG nova.compute.manager [req-475b3893-2b0c-493d-b132-1c153a0ed494 req-850d915f-10a0-4b69-bcbf-518bc493debc service nova] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Received event network-changed-25e4c4c8-9d78-4a88-9a5e-4a364ee574a5 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1087.883998] env[62753]: DEBUG nova.compute.manager [req-475b3893-2b0c-493d-b132-1c153a0ed494 req-850d915f-10a0-4b69-bcbf-518bc493debc service nova] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Refreshing instance network info cache due to event network-changed-25e4c4c8-9d78-4a88-9a5e-4a364ee574a5. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1087.884302] env[62753]: DEBUG oslo_concurrency.lockutils [req-475b3893-2b0c-493d-b132-1c153a0ed494 req-850d915f-10a0-4b69-bcbf-518bc493debc service nova] Acquiring lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1087.884468] env[62753]: DEBUG oslo_concurrency.lockutils [req-475b3893-2b0c-493d-b132-1c153a0ed494 req-850d915f-10a0-4b69-bcbf-518bc493debc service nova] Acquired lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1087.884649] env[62753]: DEBUG nova.network.neutron [req-475b3893-2b0c-493d-b132-1c153a0ed494 req-850d915f-10a0-4b69-bcbf-518bc493debc service nova] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Refreshing network info cache for port 25e4c4c8-9d78-4a88-9a5e-4a364ee574a5 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1088.241295] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b335419f-d9ce-45d3-a485-eb0db7d42f96 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.250241] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3c75d77-d2ec-4cfe-8ac7-6d8323c859bf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.284046] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56d4005b-83a6-40a8-8c68-ccf2a09c89ec {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.292342] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332807, 'name': PowerOnVM_Task} progress is 88%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.295586] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5d08ea89-7d78-416b-831c-2de1afe43533 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.311600] env[62753]: DEBUG nova.compute.provider_tree [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1088.349652] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332808, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.146911} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.349652] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1088.350211] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c65bc4b-233d-4551-b926-4448db0ce141 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.374028] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 51790f51-c464-48cc-9891-41cae5a8f63e/51790f51-c464-48cc-9891-41cae5a8f63e.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1088.374028] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4dc6ca9f-f846-44bc-a42a-00a9be0b1d59 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.398884] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1088.398884] env[62753]: value = "task-1332809" [ 1088.398884] env[62753]: _type = "Task" [ 1088.398884] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1088.409086] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332809, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.640041] env[62753]: DEBUG nova.network.neutron [req-475b3893-2b0c-493d-b132-1c153a0ed494 req-850d915f-10a0-4b69-bcbf-518bc493debc service nova] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updated VIF entry in instance network info cache for port 25e4c4c8-9d78-4a88-9a5e-4a364ee574a5. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1088.640583] env[62753]: DEBUG nova.network.neutron [req-475b3893-2b0c-493d-b132-1c153a0ed494 req-850d915f-10a0-4b69-bcbf-518bc493debc service nova] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance_info_cache with network_info: [{"id": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "address": "fa:16:3e:88:6a:1d", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25e4c4c8-9d", "ovs_interfaceid": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1088.790701] env[62753]: DEBUG oslo_vmware.api [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332807, 'name': PowerOnVM_Task, 'duration_secs': 1.38462} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1088.791261] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1088.791261] env[62753]: INFO nova.compute.manager [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Took 8.38 seconds to spawn the instance on the hypervisor. [ 1088.791453] env[62753]: DEBUG nova.compute.manager [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1088.792333] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3361d84c-90cf-4b2d-9344-12be75bfbce4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1088.814513] env[62753]: DEBUG nova.scheduler.client.report [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1088.910996] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332809, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1088.914690] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1088.914981] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1088.915227] env[62753]: DEBUG nova.compute.manager [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Going to confirm migration 4 {{(pid=62753) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1089.144320] env[62753]: DEBUG oslo_concurrency.lockutils [req-475b3893-2b0c-493d-b132-1c153a0ed494 req-850d915f-10a0-4b69-bcbf-518bc493debc service nova] Releasing lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1089.310667] env[62753]: INFO nova.compute.manager [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Took 13.17 seconds to build instance. [ 1089.319449] env[62753]: DEBUG oslo_concurrency.lockutils [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.734s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.336910] env[62753]: INFO nova.scheduler.client.report [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Deleted allocations for instance 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac [ 1089.409793] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332809, 'name': ReconfigVM_Task, 'duration_secs': 0.515328} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.410044] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 51790f51-c464-48cc-9891-41cae5a8f63e/51790f51-c464-48cc-9891-41cae5a8f63e.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1089.410718] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ee632e34-3ac6-4235-824d-b91a99edee30 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.420249] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1089.420249] env[62753]: value = "task-1332810" [ 1089.420249] env[62753]: _type = "Task" [ 1089.420249] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.430506] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332810, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1089.490892] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "refresh_cache-eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1089.491348] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquired lock "refresh_cache-eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1089.491554] env[62753]: DEBUG nova.network.neutron [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1089.491756] env[62753]: DEBUG nova.objects.instance [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lazy-loading 'info_cache' on Instance uuid eb86a978-27fa-41b2-a7e5-4b3a8ba0a152 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1089.812256] env[62753]: DEBUG oslo_concurrency.lockutils [None req-016b2e92-9684-4b50-bad3-55bda405a584 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "a2e7fa98-566e-40f1-8da7-9318de89c14c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.681s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.843499] env[62753]: DEBUG oslo_concurrency.lockutils [None req-10efa0e7-e12c-43a8-a44d-45e9b52a78b9 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.166s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1089.933548] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332810, 'name': Rename_Task, 'duration_secs': 0.152088} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1089.933786] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1089.934086] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c993c862-651c-47f9-98e2-c7c42214d434 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1089.943948] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1089.943948] env[62753]: value = "task-1332811" [ 1089.943948] env[62753]: _type = "Task" [ 1089.943948] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1089.953415] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332811, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.167334] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "32563e1f-398f-4fc9-866f-d85d9f3f27a7" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.167617] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "32563e1f-398f-4fc9-866f-d85d9f3f27a7" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.167856] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "32563e1f-398f-4fc9-866f-d85d9f3f27a7-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1090.168086] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "32563e1f-398f-4fc9-866f-d85d9f3f27a7-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1090.168278] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "32563e1f-398f-4fc9-866f-d85d9f3f27a7-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1090.170498] env[62753]: INFO nova.compute.manager [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Terminating instance [ 1090.172258] env[62753]: DEBUG nova.compute.manager [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1090.172458] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1090.173327] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9692434a-a782-48aa-8b19-0082bac18b7e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.181433] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1090.181669] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-14b0a611-4942-4f7a-807d-b83ad56c30a5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.189578] env[62753]: DEBUG oslo_vmware.api [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1090.189578] env[62753]: value = "task-1332812" [ 1090.189578] env[62753]: _type = "Task" [ 1090.189578] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.198887] env[62753]: DEBUG oslo_vmware.api [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332812, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.456018] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332811, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.694176] env[62753]: DEBUG nova.network.neutron [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Updating instance_info_cache with network_info: [{"id": "93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9", "address": "fa:16:3e:90:68:e2", "network": {"id": "c8a75326-13c0-4fdd-a608-e2cb5d049909", "bridge": "br-int", "label": "tempest-DeleteServersTestJSON-203362661-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "1d52d6eaee934be5ab0e0003df1ce316", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a10c88d7-d13f-44fd-acee-7a734eb5f56a", "external-id": "nsx-vlan-transportzone-766", "segmentation_id": 766, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap93d26d06-fe", "ovs_interfaceid": "93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1090.703139] env[62753]: DEBUG oslo_vmware.api [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332812, 'name': PowerOffVM_Task, 'duration_secs': 0.262338} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.704254] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1090.704526] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1090.704861] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e37f9daa-83f9-4abd-9a47-7d30fffcf563 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.922852] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1090.923265] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1090.923316] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Deleting the datastore file [datastore2] 32563e1f-398f-4fc9-866f-d85d9f3f27a7 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1090.923807] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ff845580-4b1d-4ee1-8d9a-540b39e167a3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1090.933014] env[62753]: DEBUG oslo_vmware.api [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1090.933014] env[62753]: value = "task-1332814" [ 1090.933014] env[62753]: _type = "Task" [ 1090.933014] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1090.943673] env[62753]: DEBUG oslo_vmware.api [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332814, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1090.954527] env[62753]: DEBUG oslo_vmware.api [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332811, 'name': PowerOnVM_Task, 'duration_secs': 0.597788} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1090.954844] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1090.955157] env[62753]: INFO nova.compute.manager [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Took 8.30 seconds to spawn the instance on the hypervisor. [ 1090.955379] env[62753]: DEBUG nova.compute.manager [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1090.956189] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eb6fc6fc-f723-4026-a5a1-4e4b04644952 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.197398] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Releasing lock "refresh_cache-eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1091.197675] env[62753]: DEBUG nova.objects.instance [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lazy-loading 'migration_context' on Instance uuid eb86a978-27fa-41b2-a7e5-4b3a8ba0a152 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1091.443733] env[62753]: DEBUG oslo_vmware.api [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332814, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.130691} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1091.443949] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1091.444201] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1091.444395] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1091.444578] env[62753]: INFO nova.compute.manager [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Took 1.27 seconds to destroy the instance on the hypervisor. [ 1091.444832] env[62753]: DEBUG oslo.service.loopingcall [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1091.445059] env[62753]: DEBUG nova.compute.manager [-] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1091.445161] env[62753]: DEBUG nova.network.neutron [-] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1091.475193] env[62753]: INFO nova.compute.manager [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Took 14.95 seconds to build instance. [ 1091.701071] env[62753]: DEBUG nova.objects.base [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Object Instance lazy-loaded attributes: info_cache,migration_context {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1091.701964] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30169956-fc46-4ddf-8f0d-40f96a78ed84 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.725601] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-6e93ddf3-9b79-4673-a479-fed80f2fafc9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1091.732445] env[62753]: DEBUG oslo_vmware.api [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1091.732445] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52462c43-b754-da5f-9de2-7322543bebce" [ 1091.732445] env[62753]: _type = "Task" [ 1091.732445] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1091.741667] env[62753]: DEBUG oslo_vmware.api [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52462c43-b754-da5f-9de2-7322543bebce, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1091.781942] env[62753]: DEBUG nova.compute.manager [req-3d9db7ff-1f33-422b-a854-fcf6cf310844 req-0ca78b72-eea5-4fe3-8418-be31b6d2111a service nova] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Received event network-vif-deleted-637f03b2-1077-4b59-98b1-a33ff9e7402b {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1091.782169] env[62753]: INFO nova.compute.manager [req-3d9db7ff-1f33-422b-a854-fcf6cf310844 req-0ca78b72-eea5-4fe3-8418-be31b6d2111a service nova] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Neutron deleted interface 637f03b2-1077-4b59-98b1-a33ff9e7402b; detaching it from the instance and deleting it from the info cache [ 1091.782351] env[62753]: DEBUG nova.network.neutron [req-3d9db7ff-1f33-422b-a854-fcf6cf310844 req-0ca78b72-eea5-4fe3-8418-be31b6d2111a service nova] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1091.977343] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5827930f-221b-4fab-9765-73863bf62333 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "51790f51-c464-48cc-9891-41cae5a8f63e" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.462s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1092.178178] env[62753]: DEBUG nova.network.neutron [-] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1092.245115] env[62753]: DEBUG oslo_vmware.api [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52462c43-b754-da5f-9de2-7322543bebce, 'name': SearchDatastore_Task, 'duration_secs': 0.014471} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1092.245367] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1092.245629] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1092.284605] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-44149c1f-9394-466b-85e3-584d52ef38a1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.300262] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5e7544f-cc36-45bf-a9fb-237395eb9401 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.328934] env[62753]: DEBUG nova.compute.manager [req-3d9db7ff-1f33-422b-a854-fcf6cf310844 req-0ca78b72-eea5-4fe3-8418-be31b6d2111a service nova] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Detach interface failed, port_id=637f03b2-1077-4b59-98b1-a33ff9e7402b, reason: Instance 32563e1f-398f-4fc9-866f-d85d9f3f27a7 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1092.343711] env[62753]: INFO nova.compute.manager [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Rescuing [ 1092.343954] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "refresh_cache-51790f51-c464-48cc-9891-41cae5a8f63e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1092.344213] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired lock "refresh_cache-51790f51-c464-48cc-9891-41cae5a8f63e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1092.344396] env[62753]: DEBUG nova.network.neutron [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1092.681261] env[62753]: INFO nova.compute.manager [-] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Took 1.24 seconds to deallocate network for instance. [ 1092.872612] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-41f5f0ec-f0f4-483a-a03c-83a30f83857a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.881067] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-17c49a5f-020c-466f-90d8-ed829efebc04 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.912719] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eca4532b-2157-415f-acbd-298d758c3b05 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.920789] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1639f7ad-1408-4c8e-b48a-462113a8dea9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1092.936222] env[62753]: DEBUG nova.compute.provider_tree [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1093.091702] env[62753]: DEBUG nova.network.neutron [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Updating instance_info_cache with network_info: [{"id": "450be489-a9db-47f3-91e7-c6673a8af916", "address": "fa:16:3e:f1:67:9b", "network": {"id": "939bb8aa-f8d4-44cb-960f-c5a5fbcf99c8", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1149502111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.11", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dd81bdb01ff4c39a4959f4af2acf61a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap450be489-a9", "ovs_interfaceid": "450be489-a9db-47f3-91e7-c6673a8af916", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1093.188158] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1093.440988] env[62753]: DEBUG nova.scheduler.client.report [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1093.597930] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Releasing lock "refresh_cache-51790f51-c464-48cc-9891-41cae5a8f63e" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1094.130751] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1094.131033] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-647828bf-65ff-4bd3-a3bc-a6f245fd723e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.139719] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1094.139719] env[62753]: value = "task-1332815" [ 1094.139719] env[62753]: _type = "Task" [ 1094.139719] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.149264] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332815, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1094.451959] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.206s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1094.454957] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.267s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1094.455218] env[62753]: DEBUG nova.objects.instance [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lazy-loading 'resources' on Instance uuid 32563e1f-398f-4fc9-866f-d85d9f3f27a7 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1094.650269] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332815, 'name': PowerOffVM_Task, 'duration_secs': 0.227517} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1094.650613] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1094.651462] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b181d2fc-0427-4c03-aaf3-3da91d1103cc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.669789] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6f128fb1-f6e3-4140-bfc2-c447526375e1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.696980] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1094.697290] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-cdc0367d-7333-473d-bf24-dbb6a9819446 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1094.704267] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1094.704267] env[62753]: value = "task-1332816" [ 1094.704267] env[62753]: _type = "Task" [ 1094.704267] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1094.711831] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332816, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.070206] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-241e8c5f-9e36-455c-8333-91834a700aeb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.081120] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8fba6df9-8335-446d-a4ed-af66beb2d627 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.112498] env[62753]: INFO nova.scheduler.client.report [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Deleted allocation for migration 6a1e13cb-e250-4311-b57b-51a150299b67 [ 1095.114060] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c952370d-e14a-4cb5-b107-7734631775b9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.125076] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f3826a0-856a-468a-b076-1370af33e7fb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.140385] env[62753]: DEBUG nova.compute.provider_tree [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1095.214625] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] VM already powered off {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1095.214867] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1095.215171] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1095.215362] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1095.215522] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1095.215801] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-1b649019-5ec8-4a76-b865-f3c69f1785f6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.225377] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1095.225563] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1095.226309] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-dc8dfcde-c679-466e-93d2-8ec9f131986b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.231559] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1095.231559] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5262bdc0-01df-09d1-54bb-b884045f7253" [ 1095.231559] env[62753]: _type = "Task" [ 1095.231559] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.239265] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5262bdc0-01df-09d1-54bb-b884045f7253, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1095.621362] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.706s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1095.643552] env[62753]: DEBUG nova.scheduler.client.report [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1095.742177] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5262bdc0-01df-09d1-54bb-b884045f7253, 'name': SearchDatastore_Task, 'duration_secs': 0.00883} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1095.742959] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-16741026-d36d-46d5-b2cc-f1130a92d96f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1095.748159] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1095.748159] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52893a8c-47f8-f713-cfc1-87411a99246f" [ 1095.748159] env[62753]: _type = "Task" [ 1095.748159] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1095.757125] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52893a8c-47f8-f713-cfc1-87411a99246f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.149050] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.694s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.167572] env[62753]: INFO nova.scheduler.client.report [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Deleted allocations for instance 32563e1f-398f-4fc9-866f-d85d9f3f27a7 [ 1096.260662] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52893a8c-47f8-f713-cfc1-87411a99246f, 'name': SearchDatastore_Task, 'duration_secs': 0.009731} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.260969] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1096.261284] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 51790f51-c464-48cc-9891-41cae5a8f63e/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk. {{(pid=62753) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1096.261596] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ce8491c0-9d02-442e-881f-78ed6f29d6e1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.270075] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1096.270075] env[62753]: value = "task-1332817" [ 1096.270075] env[62753]: _type = "Task" [ 1096.270075] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.279091] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332817, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.675924] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9d4317b2-a61d-405c-a3c4-dc25b2a95549 tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "32563e1f-398f-4fc9-866f-d85d9f3f27a7" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.508s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.780080] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332817, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.500605} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1096.780413] env[62753]: INFO nova.virt.vmwareapi.ds_util [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 51790f51-c464-48cc-9891-41cae5a8f63e/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk. [ 1096.781357] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa7c8fc5-04d3-410d-ac5f-490f15313758 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.807712] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Reconfiguring VM instance instance-0000006a to attach disk [datastore2] 51790f51-c464-48cc-9891-41cae5a8f63e/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1096.807999] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e4bf9d11-5826-47c0-976b-1972c90c609e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.827066] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1096.827066] env[62753]: value = "task-1332818" [ 1096.827066] env[62753]: _type = "Task" [ 1096.827066] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.834645] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332818, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.934506] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.934838] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.935169] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.935421] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.935657] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.938049] env[62753]: INFO nova.compute.manager [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Terminating instance [ 1096.939800] env[62753]: DEBUG nova.compute.manager [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1096.939999] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1096.940871] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-61dfad4a-ce72-4802-8e33-801075edda7b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.949550] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1096.949779] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ccf05a2c-9790-41e7-8c1e-7e450964f555 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.957014] env[62753]: DEBUG oslo_vmware.api [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1096.957014] env[62753]: value = "task-1332819" [ 1096.957014] env[62753]: _type = "Task" [ 1096.957014] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.961941] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "6e417227-895d-4576-b025-7d4a0aafa379" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.962261] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "6e417227-895d-4576-b025-7d4a0aafa379" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.962576] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "6e417227-895d-4576-b025-7d4a0aafa379-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1096.962800] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "6e417227-895d-4576-b025-7d4a0aafa379-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1096.963119] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "6e417227-895d-4576-b025-7d4a0aafa379-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1096.969894] env[62753]: DEBUG oslo_vmware.api [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332819, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1096.970352] env[62753]: INFO nova.compute.manager [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Terminating instance [ 1096.972325] env[62753]: DEBUG nova.compute.manager [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1096.972520] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1096.973347] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-db2d4efb-7da4-4e9a-9c61-9784dd046d14 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.980517] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1096.980763] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e54dc722-7613-4827-909d-91b056c10386 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1096.987798] env[62753]: DEBUG oslo_vmware.api [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1096.987798] env[62753]: value = "task-1332820" [ 1096.987798] env[62753]: _type = "Task" [ 1096.987798] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1096.996328] env[62753]: DEBUG oslo_vmware.api [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332820, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.337650] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332818, 'name': ReconfigVM_Task, 'duration_secs': 0.30111} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.338049] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Reconfigured VM instance instance-0000006a to attach disk [datastore2] 51790f51-c464-48cc-9891-41cae5a8f63e/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1097.338856] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a274b475-a176-442f-bcb9-0d2dc6fa2b8b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.365791] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-15333f0d-d26f-49fc-87e1-bdbc7d0a1e5d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.381216] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1097.381216] env[62753]: value = "task-1332821" [ 1097.381216] env[62753]: _type = "Task" [ 1097.381216] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.389321] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332821, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.467721] env[62753]: DEBUG oslo_vmware.api [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332819, 'name': PowerOffVM_Task, 'duration_secs': 0.39079} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.467996] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1097.468189] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1097.468449] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ac25ad6f-695f-440a-89f2-b8c114b1bb7a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.498128] env[62753]: DEBUG oslo_vmware.api [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332820, 'name': PowerOffVM_Task, 'duration_secs': 0.218443} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1097.498399] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1097.498573] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1097.498835] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-979f87e3-5850-499e-9461-91cf18698408 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.588214] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1097.588414] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1097.588601] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Deleting the datastore file [datastore2] 6e417227-895d-4576-b025-7d4a0aafa379 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1097.588876] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-29243fad-4465-488e-9d7c-871d89b58d18 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1097.596134] env[62753]: DEBUG oslo_vmware.api [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for the task: (returnval){ [ 1097.596134] env[62753]: value = "task-1332824" [ 1097.596134] env[62753]: _type = "Task" [ 1097.596134] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1097.604134] env[62753]: DEBUG oslo_vmware.api [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332824, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1097.892398] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332821, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.106061] env[62753]: DEBUG oslo_vmware.api [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332824, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.393155] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332821, 'name': ReconfigVM_Task, 'duration_secs': 0.699883} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.393503] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1098.393624] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6af29737-9d9d-4b5e-bd66-8daa611b97b0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.400287] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1098.400287] env[62753]: value = "task-1332825" [ 1098.400287] env[62753]: _type = "Task" [ 1098.400287] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.408343] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332825, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.455458] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1098.455688] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1098.455875] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Deleting the datastore file [datastore1] eb86a978-27fa-41b2-a7e5-4b3a8ba0a152 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1098.456249] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-14dcf256-6667-4a56-8543-dbe8b9ddf566 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.464058] env[62753]: DEBUG oslo_vmware.api [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for the task: (returnval){ [ 1098.464058] env[62753]: value = "task-1332826" [ 1098.464058] env[62753]: _type = "Task" [ 1098.464058] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1098.472548] env[62753]: DEBUG oslo_vmware.api [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332826, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1098.607995] env[62753]: DEBUG oslo_vmware.api [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Task: {'id': task-1332824, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.528134} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.608300] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1098.608496] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1098.608676] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1098.609160] env[62753]: INFO nova.compute.manager [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Took 1.64 seconds to destroy the instance on the hypervisor. [ 1098.609160] env[62753]: DEBUG oslo.service.loopingcall [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1098.609348] env[62753]: DEBUG nova.compute.manager [-] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1098.609404] env[62753]: DEBUG nova.network.neutron [-] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1098.877377] env[62753]: DEBUG nova.compute.manager [req-55c662cf-a7b9-4fd6-9a23-face65a1153c req-c5846135-21a6-4d0a-aa3c-47d6a4739243 service nova] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Received event network-vif-deleted-2254df7c-8da6-4c8d-ae9c-5b10e79dc88f {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1098.877621] env[62753]: INFO nova.compute.manager [req-55c662cf-a7b9-4fd6-9a23-face65a1153c req-c5846135-21a6-4d0a-aa3c-47d6a4739243 service nova] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Neutron deleted interface 2254df7c-8da6-4c8d-ae9c-5b10e79dc88f; detaching it from the instance and deleting it from the info cache [ 1098.877872] env[62753]: DEBUG nova.network.neutron [req-55c662cf-a7b9-4fd6-9a23-face65a1153c req-c5846135-21a6-4d0a-aa3c-47d6a4739243 service nova] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1098.910807] env[62753]: DEBUG oslo_vmware.api [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332825, 'name': PowerOnVM_Task, 'duration_secs': 0.401672} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.911339] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1098.914161] env[62753]: DEBUG nova.compute.manager [None req-b3266c58-b9f9-44e0-8a60-0bd86de9299b tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1098.914970] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9c419aea-c029-4c3c-9f9d-cbf0c8e82ff0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1098.973881] env[62753]: DEBUG oslo_vmware.api [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Task: {'id': task-1332826, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.227561} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1098.974164] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1098.974354] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1098.974544] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1098.974717] env[62753]: INFO nova.compute.manager [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Took 2.03 seconds to destroy the instance on the hypervisor. [ 1098.974993] env[62753]: DEBUG oslo.service.loopingcall [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1098.975204] env[62753]: DEBUG nova.compute.manager [-] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1098.975303] env[62753]: DEBUG nova.network.neutron [-] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1099.357913] env[62753]: DEBUG nova.network.neutron [-] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1099.380707] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-cad830d1-cd23-4c09-a239-b373e1102c27 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.390973] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ccbb400-4519-4fef-abf2-efb9b29d742e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1099.418718] env[62753]: DEBUG nova.compute.manager [req-55c662cf-a7b9-4fd6-9a23-face65a1153c req-c5846135-21a6-4d0a-aa3c-47d6a4739243 service nova] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Detach interface failed, port_id=2254df7c-8da6-4c8d-ae9c-5b10e79dc88f, reason: Instance 6e417227-895d-4576-b025-7d4a0aafa379 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1099.862596] env[62753]: INFO nova.compute.manager [-] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Took 1.25 seconds to deallocate network for instance. [ 1099.880426] env[62753]: DEBUG nova.network.neutron [-] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1100.370276] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.370276] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1100.370486] env[62753]: DEBUG nova.objects.instance [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lazy-loading 'resources' on Instance uuid 6e417227-895d-4576-b025-7d4a0aafa379 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1100.382515] env[62753]: INFO nova.compute.manager [-] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Took 1.41 seconds to deallocate network for instance. [ 1100.890055] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1100.907306] env[62753]: DEBUG nova.compute.manager [req-cf6093b2-d103-495a-81ae-e407ec79d8d9 req-b7159118-8876-4805-a3b6-630e219900ea service nova] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Received event network-vif-deleted-93d26d06-fe6d-4a32-8b36-b9c0ae8bfed9 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1100.991281] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-819d7688-7b70-4486-bc18-4bf24d0145ed {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.001724] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-edc3beb5-93a9-4106-ad75-65895f8edb91 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.034430] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d013fa43-6a7f-4be3-b377-ca38988e6c27 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.043728] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cae19bc7-f625-4bac-ad46-a6f1e2b4c993 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1101.058927] env[62753]: DEBUG nova.compute.provider_tree [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1101.562394] env[62753]: DEBUG nova.scheduler.client.report [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1101.609245] env[62753]: INFO nova.compute.manager [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Rescuing [ 1101.609513] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "refresh_cache-a2e7fa98-566e-40f1-8da7-9318de89c14c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1101.609668] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired lock "refresh_cache-a2e7fa98-566e-40f1-8da7-9318de89c14c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1101.610473] env[62753]: DEBUG nova.network.neutron [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1102.067965] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.698s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.070555] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.181s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1102.070766] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.085651] env[62753]: INFO nova.scheduler.client.report [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Deleted allocations for instance 6e417227-895d-4576-b025-7d4a0aafa379 [ 1102.087358] env[62753]: INFO nova.scheduler.client.report [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Deleted allocations for instance eb86a978-27fa-41b2-a7e5-4b3a8ba0a152 [ 1102.306755] env[62753]: DEBUG nova.network.neutron [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Updating instance_info_cache with network_info: [{"id": "dd52d78f-614d-4686-a0aa-f92598ea6932", "address": "fa:16:3e:a7:28:24", "network": {"id": "939bb8aa-f8d4-44cb-960f-c5a5fbcf99c8", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1149502111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dd81bdb01ff4c39a4959f4af2acf61a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd52d78f-61", "ovs_interfaceid": "dd52d78f-614d-4686-a0aa-f92598ea6932", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1102.598096] env[62753]: DEBUG oslo_concurrency.lockutils [None req-fd07e774-a9d2-44af-9168-9539ac2bbf9a tempest-ListServerFiltersTestJSON-82812918 tempest-ListServerFiltersTestJSON-82812918-project-member] Lock "6e417227-895d-4576-b025-7d4a0aafa379" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.636s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.599067] env[62753]: DEBUG oslo_concurrency.lockutils [None req-17978dae-8ce3-4e60-8c0f-d34262f2e04c tempest-DeleteServersTestJSON-1024143321 tempest-DeleteServersTestJSON-1024143321-project-member] Lock "eb86a978-27fa-41b2-a7e5-4b3a8ba0a152" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.664s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1102.809752] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Releasing lock "refresh_cache-a2e7fa98-566e-40f1-8da7-9318de89c14c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1103.339429] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1103.339746] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b9bb5892-9fb3-43ee-b1ed-f705db8783e3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.350417] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1103.350417] env[62753]: value = "task-1332827" [ 1103.350417] env[62753]: _type = "Task" [ 1103.350417] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.360353] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332827, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1103.860031] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332827, 'name': PowerOffVM_Task, 'duration_secs': 0.183172} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1103.860143] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1103.860920] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d20f3a0c-58d3-47f6-8677-e3ec9c3b00a7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.881731] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1f00dd8-f780-49ac-bbec-68a5ed96668d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.928578] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1103.928900] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-175f9b72-1aa5-4c66-8e8b-4050ba021400 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1103.937629] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1103.937629] env[62753]: value = "task-1332829" [ 1103.937629] env[62753]: _type = "Task" [ 1103.937629] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1103.946155] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332829, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.449030] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] VM already powered off {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1104.449354] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1104.449395] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1104.449516] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1104.449693] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1104.449945] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2ad82973-5268-485a-9829-4809f62cc47c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.459465] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1104.459664] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1104.460432] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-bf7e203a-9b2b-47c4-a8a5-0d22294efc6f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.466590] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1104.466590] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52eb5bf2-0e1a-edb2-68cf-a5e9576bda3b" [ 1104.466590] env[62753]: _type = "Task" [ 1104.466590] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.474635] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52eb5bf2-0e1a-edb2-68cf-a5e9576bda3b, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.752103] env[62753]: DEBUG oslo_concurrency.lockutils [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "01d9da0b-f4e5-474c-aab2-a0f4f0c99994" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.752103] env[62753]: DEBUG oslo_concurrency.lockutils [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "01d9da0b-f4e5-474c-aab2-a0f4f0c99994" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.752103] env[62753]: DEBUG oslo_concurrency.lockutils [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "01d9da0b-f4e5-474c-aab2-a0f4f0c99994-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1104.752103] env[62753]: DEBUG oslo_concurrency.lockutils [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "01d9da0b-f4e5-474c-aab2-a0f4f0c99994-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1104.752103] env[62753]: DEBUG oslo_concurrency.lockutils [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "01d9da0b-f4e5-474c-aab2-a0f4f0c99994-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1104.755034] env[62753]: INFO nova.compute.manager [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Terminating instance [ 1104.756449] env[62753]: DEBUG nova.compute.manager [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1104.756656] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1104.757559] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff8f845b-6dd9-49e9-a254-77aa852b6cd8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.768553] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1104.768903] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6897f0a1-14da-4927-b91c-4ad890cda47c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.775619] env[62753]: DEBUG oslo_vmware.api [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1104.775619] env[62753]: value = "task-1332830" [ 1104.775619] env[62753]: _type = "Task" [ 1104.775619] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.785522] env[62753]: DEBUG oslo_vmware.api [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332830, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1104.977746] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52eb5bf2-0e1a-edb2-68cf-a5e9576bda3b, 'name': SearchDatastore_Task, 'duration_secs': 0.009929} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1104.978581] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-ad258f77-a87e-4ced-9fee-df19247c9071 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1104.985054] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1104.985054] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525b33cc-77f8-4c83-f15a-313740958aa9" [ 1104.985054] env[62753]: _type = "Task" [ 1104.985054] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1104.993321] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525b33cc-77f8-4c83-f15a-313740958aa9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.290086] env[62753]: DEBUG oslo_vmware.api [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332830, 'name': PowerOffVM_Task, 'duration_secs': 0.383792} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.290715] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1105.290941] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1105.291805] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3c241de8-575c-4b05-82c3-2b2707b5d5f6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.363122] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1105.363399] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1105.363592] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Deleting the datastore file [datastore1] 01d9da0b-f4e5-474c-aab2-a0f4f0c99994 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1105.363876] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-cc4d3fb8-b75f-47d0-ad8b-fac424d8a0f6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.371900] env[62753]: DEBUG oslo_vmware.api [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1105.371900] env[62753]: value = "task-1332832" [ 1105.371900] env[62753]: _type = "Task" [ 1105.371900] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.382014] env[62753]: DEBUG oslo_vmware.api [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332832, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.500929] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525b33cc-77f8-4c83-f15a-313740958aa9, 'name': SearchDatastore_Task, 'duration_secs': 0.009444} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.500929] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1105.500929] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] a2e7fa98-566e-40f1-8da7-9318de89c14c/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk. {{(pid=62753) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1105.500929] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-3d047daf-f7c2-41b9-a95d-023ff8d3fbe8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1105.511577] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1105.511577] env[62753]: value = "task-1332833" [ 1105.511577] env[62753]: _type = "Task" [ 1105.511577] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1105.522092] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332833, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1105.890030] env[62753]: DEBUG oslo_vmware.api [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332832, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.15324} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1105.890438] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1105.890685] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1105.890901] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1105.891264] env[62753]: INFO nova.compute.manager [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Took 1.13 seconds to destroy the instance on the hypervisor. [ 1105.891530] env[62753]: DEBUG oslo.service.loopingcall [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1105.891779] env[62753]: DEBUG nova.compute.manager [-] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1105.891883] env[62753]: DEBUG nova.network.neutron [-] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1106.023092] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332833, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.507379} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.023853] env[62753]: INFO nova.virt.vmwareapi.ds_util [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] a2e7fa98-566e-40f1-8da7-9318de89c14c/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk. [ 1106.024732] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-25519ea0-e3b3-4f30-bc8c-e7470bddb401 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.054767] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Reconfiguring VM instance instance-00000069 to attach disk [datastore2] a2e7fa98-566e-40f1-8da7-9318de89c14c/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1106.055279] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-667a4e25-f0a0-420b-8823-1ca45b44f7ae {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.079092] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1106.079092] env[62753]: value = "task-1332834" [ 1106.079092] env[62753]: _type = "Task" [ 1106.079092] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.089436] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332834, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1106.542824] env[62753]: DEBUG nova.compute.manager [req-cb7f0be3-438a-4082-8e6c-68f9790736c7 req-5d376eae-28ba-4dd9-b0ea-f159582eadaa service nova] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Received event network-vif-deleted-d3934283-dc65-4a50-8b00-c466bb372792 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1106.544363] env[62753]: INFO nova.compute.manager [req-cb7f0be3-438a-4082-8e6c-68f9790736c7 req-5d376eae-28ba-4dd9-b0ea-f159582eadaa service nova] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Neutron deleted interface d3934283-dc65-4a50-8b00-c466bb372792; detaching it from the instance and deleting it from the info cache [ 1106.544592] env[62753]: DEBUG nova.network.neutron [req-cb7f0be3-438a-4082-8e6c-68f9790736c7 req-5d376eae-28ba-4dd9-b0ea-f159582eadaa service nova] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1106.594363] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332834, 'name': ReconfigVM_Task, 'duration_secs': 0.359136} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1106.594714] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Reconfigured VM instance instance-00000069 to attach disk [datastore2] a2e7fa98-566e-40f1-8da7-9318de89c14c/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1106.595656] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf4e099b-c22c-4171-9c64-d4a40d06b71f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.629691] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-39d98dd5-237b-4d2d-b3cc-02689e573525 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1106.649685] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1106.649685] env[62753]: value = "task-1332835" [ 1106.649685] env[62753]: _type = "Task" [ 1106.649685] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1106.658794] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332835, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.014407] env[62753]: DEBUG nova.network.neutron [-] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1107.047724] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-e8a64e04-3404-420f-82ff-3aee175af93b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.058281] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7fea63a-a169-4c57-9661-bb43c61ebcad {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1107.087533] env[62753]: DEBUG nova.compute.manager [req-cb7f0be3-438a-4082-8e6c-68f9790736c7 req-5d376eae-28ba-4dd9-b0ea-f159582eadaa service nova] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Detach interface failed, port_id=d3934283-dc65-4a50-8b00-c466bb372792, reason: Instance 01d9da0b-f4e5-474c-aab2-a0f4f0c99994 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1107.160331] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332835, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1107.520398] env[62753]: INFO nova.compute.manager [-] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Took 1.63 seconds to deallocate network for instance. [ 1107.662107] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332835, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.027210] env[62753]: DEBUG oslo_concurrency.lockutils [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1108.028076] env[62753]: DEBUG oslo_concurrency.lockutils [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1108.029490] env[62753]: DEBUG nova.objects.instance [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lazy-loading 'resources' on Instance uuid 01d9da0b-f4e5-474c-aab2-a0f4f0c99994 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1108.161384] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332835, 'name': ReconfigVM_Task, 'duration_secs': 1.169458} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.161678] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1108.161955] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d0d8c826-57ff-4bd0-a017-4290aecdaf03 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.169029] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1108.169029] env[62753]: value = "task-1332839" [ 1108.169029] env[62753]: _type = "Task" [ 1108.169029] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1108.177513] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332839, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1108.619510] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b959f871-50cc-41ba-8b1b-b35106af2eca {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.629244] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b757128b-61d0-4e08-8916-dcdc6f4d0e19 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.660245] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9fb56c3-d091-4f4b-b5a9-fee4e7948802 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.667602] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e415a7b6-af9b-4bdb-be7a-05dff228147e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1108.683305] env[62753]: DEBUG nova.compute.provider_tree [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1108.687787] env[62753]: DEBUG oslo_vmware.api [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332839, 'name': PowerOnVM_Task, 'duration_secs': 0.416488} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1108.688039] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1108.690466] env[62753]: DEBUG nova.compute.manager [None req-cd0b3119-4cbe-4bb6-a357-c6d5ff003e72 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1108.692120] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30929dd4-1338-4641-83a4-afae604561df {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1109.187244] env[62753]: DEBUG nova.scheduler.client.report [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1109.692925] env[62753]: DEBUG oslo_concurrency.lockutils [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.665s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1109.710309] env[62753]: INFO nova.compute.manager [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Unrescuing [ 1109.710520] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "refresh_cache-a2e7fa98-566e-40f1-8da7-9318de89c14c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1109.710741] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired lock "refresh_cache-a2e7fa98-566e-40f1-8da7-9318de89c14c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1109.711013] env[62753]: DEBUG nova.network.neutron [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1109.713166] env[62753]: INFO nova.scheduler.client.report [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Deleted allocations for instance 01d9da0b-f4e5-474c-aab2-a0f4f0c99994 [ 1110.219968] env[62753]: DEBUG oslo_concurrency.lockutils [None req-79a62b8c-69cf-4610-962d-da2b99c64c5a tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "01d9da0b-f4e5-474c-aab2-a0f4f0c99994" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.468s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1110.413535] env[62753]: DEBUG nova.network.neutron [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Updating instance_info_cache with network_info: [{"id": "dd52d78f-614d-4686-a0aa-f92598ea6932", "address": "fa:16:3e:a7:28:24", "network": {"id": "939bb8aa-f8d4-44cb-960f-c5a5fbcf99c8", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1149502111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dd81bdb01ff4c39a4959f4af2acf61a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapdd52d78f-61", "ovs_interfaceid": "dd52d78f-614d-4686-a0aa-f92598ea6932", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1110.753840] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "04fd0022-f20a-4217-9e47-5381635b17a6" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1110.754195] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "04fd0022-f20a-4217-9e47-5381635b17a6" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1110.917205] env[62753]: DEBUG oslo_concurrency.lockutils [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Releasing lock "refresh_cache-a2e7fa98-566e-40f1-8da7-9318de89c14c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1110.917905] env[62753]: DEBUG nova.objects.instance [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lazy-loading 'flavor' on Instance uuid a2e7fa98-566e-40f1-8da7-9318de89c14c {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1111.257153] env[62753]: INFO nova.compute.manager [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Detaching volume e93861cc-b07a-4254-b33d-0814321eef34 [ 1111.286969] env[62753]: INFO nova.virt.block_device [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Attempting to driver detach volume e93861cc-b07a-4254-b33d-0814321eef34 from mountpoint /dev/sdb [ 1111.287197] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Volume detach. Driver type: vmdk {{(pid=62753) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1111.287385] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284671', 'volume_id': 'e93861cc-b07a-4254-b33d-0814321eef34', 'name': 'volume-e93861cc-b07a-4254-b33d-0814321eef34', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '04fd0022-f20a-4217-9e47-5381635b17a6', 'attached_at': '', 'detached_at': '', 'volume_id': 'e93861cc-b07a-4254-b33d-0814321eef34', 'serial': 'e93861cc-b07a-4254-b33d-0814321eef34'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1111.288271] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7d068849-44b6-442e-8a62-feaa5415f34c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.309786] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b0a5d04-873c-44d4-9d3b-0501df612f41 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.316401] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5be07b2-66c8-4ee4-9d89-cf87136f6b05 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.336996] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e75432f4-54ca-4af9-9a85-effb0aef45a4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.351129] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] The volume has not been displaced from its original location: [datastore1] volume-e93861cc-b07a-4254-b33d-0814321eef34/volume-e93861cc-b07a-4254-b33d-0814321eef34.vmdk. No consolidation needed. {{(pid=62753) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1111.356289] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Reconfiguring VM instance instance-00000058 to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1111.356556] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b1f2383e-c136-4781-ad80-746fc54b8c78 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.373592] env[62753]: DEBUG oslo_vmware.api [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1111.373592] env[62753]: value = "task-1332841" [ 1111.373592] env[62753]: _type = "Task" [ 1111.373592] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.381050] env[62753]: DEBUG oslo_vmware.api [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332841, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.425052] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bcea1916-52bb-45d0-bae6-a7269168c197 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.446513] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1111.446851] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-909b4188-967f-4d99-8632-a4398342f32b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.453983] env[62753]: DEBUG oslo_vmware.api [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1111.453983] env[62753]: value = "task-1332842" [ 1111.453983] env[62753]: _type = "Task" [ 1111.453983] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.461787] env[62753]: DEBUG oslo_vmware.api [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332842, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.883485] env[62753]: DEBUG oslo_vmware.api [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332841, 'name': ReconfigVM_Task, 'duration_secs': 0.288897} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.883793] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Reconfigured VM instance instance-00000058 to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1111.888738] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c006c63e-b84c-423e-abe3-4a9bb4d845eb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.903605] env[62753]: DEBUG oslo_vmware.api [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1111.903605] env[62753]: value = "task-1332843" [ 1111.903605] env[62753]: _type = "Task" [ 1111.903605] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1111.911068] env[62753]: DEBUG oslo_vmware.api [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332843, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1111.963161] env[62753]: DEBUG oslo_vmware.api [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332842, 'name': PowerOffVM_Task, 'duration_secs': 0.210095} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1111.963419] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1111.968647] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Reconfiguring VM instance instance-00000069 to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1111.968926] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2806a044-9893-490b-8313-03c6e8d56779 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1111.986757] env[62753]: DEBUG oslo_vmware.api [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1111.986757] env[62753]: value = "task-1332844" [ 1111.986757] env[62753]: _type = "Task" [ 1111.986757] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.000311] env[62753]: DEBUG oslo_vmware.api [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332844, 'name': ReconfigVM_Task} progress is 10%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.413561] env[62753]: DEBUG oslo_vmware.api [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332843, 'name': ReconfigVM_Task, 'duration_secs': 0.128668} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.413883] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284671', 'volume_id': 'e93861cc-b07a-4254-b33d-0814321eef34', 'name': 'volume-e93861cc-b07a-4254-b33d-0814321eef34', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '04fd0022-f20a-4217-9e47-5381635b17a6', 'attached_at': '', 'detached_at': '', 'volume_id': 'e93861cc-b07a-4254-b33d-0814321eef34', 'serial': 'e93861cc-b07a-4254-b33d-0814321eef34'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1112.496658] env[62753]: DEBUG oslo_vmware.api [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332844, 'name': ReconfigVM_Task, 'duration_secs': 0.218903} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1112.497017] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Reconfigured VM instance instance-00000069 to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1112.497278] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1112.497601] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c11038f9-ceb2-4857-8efe-f9d8897b4ce6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1112.504917] env[62753]: DEBUG oslo_vmware.api [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1112.504917] env[62753]: value = "task-1332845" [ 1112.504917] env[62753]: _type = "Task" [ 1112.504917] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1112.513641] env[62753]: DEBUG oslo_vmware.api [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332845, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1112.956903] env[62753]: DEBUG nova.objects.instance [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lazy-loading 'flavor' on Instance uuid 04fd0022-f20a-4217-9e47-5381635b17a6 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1113.015304] env[62753]: DEBUG oslo_vmware.api [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332845, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1113.514180] env[62753]: DEBUG oslo_vmware.api [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332845, 'name': PowerOnVM_Task, 'duration_secs': 0.516541} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1113.514450] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1113.514671] env[62753]: DEBUG nova.compute.manager [None req-a9740937-ada8-4d62-aef5-d2cdc2dd0dee tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1113.515469] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb6e63a5-2b46-4675-b315-cdd6b0af1490 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1113.965263] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0036e1a3-d614-454e-a555-09991859c6c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "04fd0022-f20a-4217-9e47-5381635b17a6" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.211s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.978769] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "04fd0022-f20a-4217-9e47-5381635b17a6" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.979092] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "04fd0022-f20a-4217-9e47-5381635b17a6" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.979275] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "04fd0022-f20a-4217-9e47-5381635b17a6-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1114.979507] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "04fd0022-f20a-4217-9e47-5381635b17a6-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1114.979692] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "04fd0022-f20a-4217-9e47-5381635b17a6-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1114.982179] env[62753]: INFO nova.compute.manager [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Terminating instance [ 1114.984020] env[62753]: DEBUG nova.compute.manager [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1114.984228] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1114.985078] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcd3d4f5-d1f2-47d7-ab0c-3e0464743187 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1114.993482] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1114.993662] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ebc8a127-042b-4a00-bc8f-9f55dc861d98 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.001612] env[62753]: DEBUG oslo_vmware.api [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1115.001612] env[62753]: value = "task-1332847" [ 1115.001612] env[62753]: _type = "Task" [ 1115.001612] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.010459] env[62753]: DEBUG oslo_vmware.api [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332847, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1115.512721] env[62753]: DEBUG oslo_vmware.api [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332847, 'name': PowerOffVM_Task, 'duration_secs': 0.233143} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1115.513019] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1115.513304] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1115.513773] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-ad5c963f-3ff1-44b0-94f5-57c58d32be56 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.602673] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1115.602928] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1115.603138] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Deleting the datastore file [datastore2] 04fd0022-f20a-4217-9e47-5381635b17a6 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1115.603431] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f3c5ffd3-82fd-45cf-83e7-2da78e0fc8a3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1115.611632] env[62753]: DEBUG oslo_vmware.api [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1115.611632] env[62753]: value = "task-1332849" [ 1115.611632] env[62753]: _type = "Task" [ 1115.611632] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1115.620541] env[62753]: DEBUG oslo_vmware.api [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332849, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1116.121359] env[62753]: DEBUG oslo_vmware.api [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332849, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.265983} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1116.121732] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1116.121812] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1116.121995] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1116.122192] env[62753]: INFO nova.compute.manager [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Took 1.14 seconds to destroy the instance on the hypervisor. [ 1116.122442] env[62753]: DEBUG oslo.service.loopingcall [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1116.122643] env[62753]: DEBUG nova.compute.manager [-] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1116.122740] env[62753]: DEBUG nova.network.neutron [-] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1116.611926] env[62753]: DEBUG nova.compute.manager [req-d9dc7cc9-2950-4d19-b526-8d001bdce5dc req-8427a6f1-4848-46e6-9dcc-7c0e5290fa01 service nova] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Received event network-vif-deleted-011eae16-b648-4e2e-abfb-e424e05e41c2 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1116.612073] env[62753]: INFO nova.compute.manager [req-d9dc7cc9-2950-4d19-b526-8d001bdce5dc req-8427a6f1-4848-46e6-9dcc-7c0e5290fa01 service nova] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Neutron deleted interface 011eae16-b648-4e2e-abfb-e424e05e41c2; detaching it from the instance and deleting it from the info cache [ 1116.612276] env[62753]: DEBUG nova.network.neutron [req-d9dc7cc9-2950-4d19-b526-8d001bdce5dc req-8427a6f1-4848-46e6-9dcc-7c0e5290fa01 service nova] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1116.882899] env[62753]: DEBUG oslo_concurrency.lockutils [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1116.883279] env[62753]: DEBUG oslo_concurrency.lockutils [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.084384] env[62753]: DEBUG nova.network.neutron [-] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1117.114748] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-1991b630-88dd-4461-b0d9-86e437f5735c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.127189] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-537b8995-6bcb-4c68-b2d1-aa0a065553b5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1117.155784] env[62753]: DEBUG nova.compute.manager [req-d9dc7cc9-2950-4d19-b526-8d001bdce5dc req-8427a6f1-4848-46e6-9dcc-7c0e5290fa01 service nova] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Detach interface failed, port_id=011eae16-b648-4e2e-abfb-e424e05e41c2, reason: Instance 04fd0022-f20a-4217-9e47-5381635b17a6 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1117.385710] env[62753]: DEBUG nova.compute.manager [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1117.562682] env[62753]: DEBUG oslo_concurrency.lockutils [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.562935] env[62753]: DEBUG oslo_concurrency.lockutils [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.587012] env[62753]: INFO nova.compute.manager [-] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Took 1.46 seconds to deallocate network for instance. [ 1117.908830] env[62753]: DEBUG oslo_concurrency.lockutils [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1117.909108] env[62753]: DEBUG oslo_concurrency.lockutils [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1117.910762] env[62753]: INFO nova.compute.claims [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1118.066805] env[62753]: DEBUG nova.compute.utils [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1118.093361] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1118.570047] env[62753]: DEBUG oslo_concurrency.lockutils [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.007s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1119.007032] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d82c009-7ed3-42ac-a1df-47b0592d9f17 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.015572] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47864a16-a5c3-4931-b498-83f11d7dd6b4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.047234] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64a1e17c-bf5e-4c1e-bcfa-78fb5eb40c10 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.055681] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-316e1c5a-9429-4efd-b5ca-457ec1453a61 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.069102] env[62753]: DEBUG nova.compute.provider_tree [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1119.572870] env[62753]: DEBUG nova.scheduler.client.report [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1119.633123] env[62753]: DEBUG oslo_concurrency.lockutils [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1119.633552] env[62753]: DEBUG oslo_concurrency.lockutils [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1119.633980] env[62753]: INFO nova.compute.manager [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Attaching volume ea8a3d93-7c93-4b76-9194-d4539fc86180 to /dev/sdb [ 1119.666352] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2d5a0bfa-950c-4e35-9ae9-246e38b77b67 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.675256] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ffb8bc2-c732-4d0e-9eb2-1a898dc112ec {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1119.689264] env[62753]: DEBUG nova.virt.block_device [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating existing volume attachment record: c8196ba5-ec15-49ab-8b5e-d844cd70ea02 {{(pid=62753) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1120.077886] env[62753]: DEBUG oslo_concurrency.lockutils [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.169s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1120.078522] env[62753]: DEBUG nova.compute.manager [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1120.081433] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.988s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1120.081603] env[62753]: DEBUG nova.objects.instance [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lazy-loading 'resources' on Instance uuid 04fd0022-f20a-4217-9e47-5381635b17a6 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1120.584836] env[62753]: DEBUG nova.compute.utils [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1120.588947] env[62753]: DEBUG nova.compute.manager [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1120.589136] env[62753]: DEBUG nova.network.neutron [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1120.636624] env[62753]: DEBUG nova.policy [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '2a7f23e0237349bdbebc6135191ffe9c', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '2dd81bdb01ff4c39a4959f4af2acf61a', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1120.672867] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6f453fa-f0cc-4a8c-b337-520af506b166 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.681648] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa63ed2-54a3-4853-9a5f-48e2c4219912 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.714077] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dcbc418b-6420-4526-86a1-dbe75f37b455 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.722498] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-103a825f-946b-48f1-b6e0-cbf07bbfefd7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1120.737099] env[62753]: DEBUG nova.compute.provider_tree [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1120.916517] env[62753]: DEBUG nova.network.neutron [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Successfully created port: 8600276a-3346-44f1-a760-df6caca2e51e {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1121.089615] env[62753]: DEBUG nova.compute.manager [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1121.239946] env[62753]: DEBUG nova.scheduler.client.report [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1121.744632] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.663s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1121.767350] env[62753]: INFO nova.scheduler.client.report [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Deleted allocations for instance 04fd0022-f20a-4217-9e47-5381635b17a6 [ 1122.101422] env[62753]: DEBUG nova.compute.manager [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1122.126997] env[62753]: DEBUG nova.virt.hardware [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1122.127611] env[62753]: DEBUG nova.virt.hardware [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1122.127611] env[62753]: DEBUG nova.virt.hardware [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1122.127611] env[62753]: DEBUG nova.virt.hardware [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1122.127796] env[62753]: DEBUG nova.virt.hardware [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1122.127955] env[62753]: DEBUG nova.virt.hardware [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1122.128191] env[62753]: DEBUG nova.virt.hardware [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1122.128357] env[62753]: DEBUG nova.virt.hardware [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1122.128564] env[62753]: DEBUG nova.virt.hardware [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1122.128759] env[62753]: DEBUG nova.virt.hardware [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1122.128944] env[62753]: DEBUG nova.virt.hardware [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1122.129847] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27efcb68-2ebe-41ad-8ff1-5d265ea41b1e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.138161] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1af160bf-e66c-4c33-95f0-0986f57a3c5c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1122.275678] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c67b8e81-79d7-4145-86f1-85ffe0ebef48 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "04fd0022-f20a-4217-9e47-5381635b17a6" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 7.296s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.370244] env[62753]: DEBUG nova.compute.manager [req-afe6d53b-211b-410d-b9d2-0e1ebe8ad3be req-18fda3ec-1fc6-416d-a586-c5e72cb45d38 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Received event network-vif-plugged-8600276a-3346-44f1-a760-df6caca2e51e {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1122.370467] env[62753]: DEBUG oslo_concurrency.lockutils [req-afe6d53b-211b-410d-b9d2-0e1ebe8ad3be req-18fda3ec-1fc6-416d-a586-c5e72cb45d38 service nova] Acquiring lock "ec0b07b8-1cb9-4911-8b03-efb466a21888-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1122.370686] env[62753]: DEBUG oslo_concurrency.lockutils [req-afe6d53b-211b-410d-b9d2-0e1ebe8ad3be req-18fda3ec-1fc6-416d-a586-c5e72cb45d38 service nova] Lock "ec0b07b8-1cb9-4911-8b03-efb466a21888-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1122.370855] env[62753]: DEBUG oslo_concurrency.lockutils [req-afe6d53b-211b-410d-b9d2-0e1ebe8ad3be req-18fda3ec-1fc6-416d-a586-c5e72cb45d38 service nova] Lock "ec0b07b8-1cb9-4911-8b03-efb466a21888-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1122.371131] env[62753]: DEBUG nova.compute.manager [req-afe6d53b-211b-410d-b9d2-0e1ebe8ad3be req-18fda3ec-1fc6-416d-a586-c5e72cb45d38 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] No waiting events found dispatching network-vif-plugged-8600276a-3346-44f1-a760-df6caca2e51e {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1122.371343] env[62753]: WARNING nova.compute.manager [req-afe6d53b-211b-410d-b9d2-0e1ebe8ad3be req-18fda3ec-1fc6-416d-a586-c5e72cb45d38 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Received unexpected event network-vif-plugged-8600276a-3346-44f1-a760-df6caca2e51e for instance with vm_state building and task_state spawning. [ 1122.458140] env[62753]: DEBUG nova.network.neutron [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Successfully updated port: 8600276a-3346-44f1-a760-df6caca2e51e {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1122.960857] env[62753]: DEBUG oslo_concurrency.lockutils [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1122.961148] env[62753]: DEBUG oslo_concurrency.lockutils [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1122.961202] env[62753]: DEBUG nova.network.neutron [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1123.491947] env[62753]: DEBUG nova.network.neutron [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1123.620847] env[62753]: DEBUG nova.network.neutron [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Updating instance_info_cache with network_info: [{"id": "8600276a-3346-44f1-a760-df6caca2e51e", "address": "fa:16:3e:bf:ad:12", "network": {"id": "939bb8aa-f8d4-44cb-960f-c5a5fbcf99c8", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1149502111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dd81bdb01ff4c39a4959f4af2acf61a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8600276a-33", "ovs_interfaceid": "8600276a-3346-44f1-a760-df6caca2e51e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1124.068332] env[62753]: DEBUG oslo_concurrency.lockutils [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "21e6cf03-e528-4289-8813-86020a4e8d2a" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1124.068665] env[62753]: DEBUG oslo_concurrency.lockutils [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "21e6cf03-e528-4289-8813-86020a4e8d2a" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1124.123315] env[62753]: DEBUG oslo_concurrency.lockutils [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Releasing lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1124.123646] env[62753]: DEBUG nova.compute.manager [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Instance network_info: |[{"id": "8600276a-3346-44f1-a760-df6caca2e51e", "address": "fa:16:3e:bf:ad:12", "network": {"id": "939bb8aa-f8d4-44cb-960f-c5a5fbcf99c8", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1149502111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dd81bdb01ff4c39a4959f4af2acf61a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8600276a-33", "ovs_interfaceid": "8600276a-3346-44f1-a760-df6caca2e51e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1124.124088] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:bf:ad:12', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '8600276a-3346-44f1-a760-df6caca2e51e', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1124.132076] env[62753]: DEBUG oslo.service.loopingcall [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1124.132243] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1124.132498] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-013b7292-1cd6-4365-b135-81e7f7b51053 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.154094] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1124.154094] env[62753]: value = "task-1332858" [ 1124.154094] env[62753]: _type = "Task" [ 1124.154094] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.162008] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332858, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.235441] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Volume attach. Driver type: vmdk {{(pid=62753) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1124.235733] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284694', 'volume_id': 'ea8a3d93-7c93-4b76-9194-d4539fc86180', 'name': 'volume-ea8a3d93-7c93-4b76-9194-d4539fc86180', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca01cd3f-7f0e-4c61-a704-74f9945792b4', 'attached_at': '', 'detached_at': '', 'volume_id': 'ea8a3d93-7c93-4b76-9194-d4539fc86180', 'serial': 'ea8a3d93-7c93-4b76-9194-d4539fc86180'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1124.236632] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-576a74ed-ce9a-4936-9f2e-8a5e8afe9530 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.253283] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5e9b9a4-23d4-49c2-a4f0-1306f84dbe33 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.277307] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Reconfiguring VM instance instance-00000067 to attach disk [datastore2] volume-ea8a3d93-7c93-4b76-9194-d4539fc86180/volume-ea8a3d93-7c93-4b76-9194-d4539fc86180.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1124.277572] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-21e0070d-541a-437b-b448-1651e5a84480 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.295768] env[62753]: DEBUG oslo_vmware.api [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1124.295768] env[62753]: value = "task-1332859" [ 1124.295768] env[62753]: _type = "Task" [ 1124.295768] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.304304] env[62753]: DEBUG oslo_vmware.api [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332859, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.400689] env[62753]: DEBUG nova.compute.manager [req-2892a052-e483-4a5e-8206-3b8cb84b23e5 req-8b4d20ee-1924-47df-9c28-d9db1d434e74 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Received event network-changed-8600276a-3346-44f1-a760-df6caca2e51e {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1124.400924] env[62753]: DEBUG nova.compute.manager [req-2892a052-e483-4a5e-8206-3b8cb84b23e5 req-8b4d20ee-1924-47df-9c28-d9db1d434e74 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Refreshing instance network info cache due to event network-changed-8600276a-3346-44f1-a760-df6caca2e51e. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1124.401163] env[62753]: DEBUG oslo_concurrency.lockutils [req-2892a052-e483-4a5e-8206-3b8cb84b23e5 req-8b4d20ee-1924-47df-9c28-d9db1d434e74 service nova] Acquiring lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1124.401314] env[62753]: DEBUG oslo_concurrency.lockutils [req-2892a052-e483-4a5e-8206-3b8cb84b23e5 req-8b4d20ee-1924-47df-9c28-d9db1d434e74 service nova] Acquired lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1124.401477] env[62753]: DEBUG nova.network.neutron [req-2892a052-e483-4a5e-8206-3b8cb84b23e5 req-8b4d20ee-1924-47df-9c28-d9db1d434e74 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Refreshing network info cache for port 8600276a-3346-44f1-a760-df6caca2e51e {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1124.571831] env[62753]: DEBUG nova.compute.manager [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1124.664464] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332858, 'name': CreateVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1124.806911] env[62753]: DEBUG oslo_vmware.api [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332859, 'name': ReconfigVM_Task, 'duration_secs': 0.336618} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1124.807173] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Reconfigured VM instance instance-00000067 to attach disk [datastore2] volume-ea8a3d93-7c93-4b76-9194-d4539fc86180/volume-ea8a3d93-7c93-4b76-9194-d4539fc86180.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1124.812103] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6a848036-ee74-4fb8-bcc8-5a42d2b1730f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1124.827853] env[62753]: DEBUG oslo_vmware.api [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1124.827853] env[62753]: value = "task-1332860" [ 1124.827853] env[62753]: _type = "Task" [ 1124.827853] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1124.836245] env[62753]: DEBUG oslo_vmware.api [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332860, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.010418] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1125.097952] env[62753]: DEBUG oslo_concurrency.lockutils [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.098325] env[62753]: DEBUG oslo_concurrency.lockutils [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.099803] env[62753]: INFO nova.compute.claims [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1125.115553] env[62753]: DEBUG nova.network.neutron [req-2892a052-e483-4a5e-8206-3b8cb84b23e5 req-8b4d20ee-1924-47df-9c28-d9db1d434e74 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Updated VIF entry in instance network info cache for port 8600276a-3346-44f1-a760-df6caca2e51e. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1125.116806] env[62753]: DEBUG nova.network.neutron [req-2892a052-e483-4a5e-8206-3b8cb84b23e5 req-8b4d20ee-1924-47df-9c28-d9db1d434e74 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Updating instance_info_cache with network_info: [{"id": "8600276a-3346-44f1-a760-df6caca2e51e", "address": "fa:16:3e:bf:ad:12", "network": {"id": "939bb8aa-f8d4-44cb-960f-c5a5fbcf99c8", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1149502111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dd81bdb01ff4c39a4959f4af2acf61a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8600276a-33", "ovs_interfaceid": "8600276a-3346-44f1-a760-df6caca2e51e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1125.166964] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332858, 'name': CreateVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.338820] env[62753]: DEBUG oslo_vmware.api [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332860, 'name': ReconfigVM_Task, 'duration_secs': 0.146139} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.339212] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284694', 'volume_id': 'ea8a3d93-7c93-4b76-9194-d4539fc86180', 'name': 'volume-ea8a3d93-7c93-4b76-9194-d4539fc86180', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ca01cd3f-7f0e-4c61-a704-74f9945792b4', 'attached_at': '', 'detached_at': '', 'volume_id': 'ea8a3d93-7c93-4b76-9194-d4539fc86180', 'serial': 'ea8a3d93-7c93-4b76-9194-d4539fc86180'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1125.618041] env[62753]: DEBUG oslo_concurrency.lockutils [req-2892a052-e483-4a5e-8206-3b8cb84b23e5 req-8b4d20ee-1924-47df-9c28-d9db1d434e74 service nova] Releasing lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1125.665154] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332858, 'name': CreateVM_Task, 'duration_secs': 1.338209} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1125.665321] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1125.666031] env[62753]: DEBUG oslo_concurrency.lockutils [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1125.666213] env[62753]: DEBUG oslo_concurrency.lockutils [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1125.666581] env[62753]: DEBUG oslo_concurrency.lockutils [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1125.666834] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8fd0de4b-626e-4915-95bd-8c3112873f16 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.671787] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1125.671787] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52184110-ada7-4bea-c2cd-c457009629d2" [ 1125.671787] env[62753]: _type = "Task" [ 1125.671787] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1125.679164] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52184110-ada7-4bea-c2cd-c457009629d2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1125.885457] env[62753]: DEBUG oslo_concurrency.lockutils [None req-067cb155-e3fe-47a2-9732-761c3529d6a7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "22cea165-cfde-403f-a0e2-82a86f5baa51" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1125.885786] env[62753]: DEBUG oslo_concurrency.lockutils [None req-067cb155-e3fe-47a2-9732-761c3529d6a7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1125.886010] env[62753]: DEBUG nova.compute.manager [None req-067cb155-e3fe-47a2-9732-761c3529d6a7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1125.887250] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-78101a03-8f1c-4030-abcb-e8d4f1bc405e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1125.896196] env[62753]: DEBUG nova.compute.manager [None req-067cb155-e3fe-47a2-9732-761c3529d6a7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62753) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1125.896708] env[62753]: DEBUG nova.objects.instance [None req-067cb155-e3fe-47a2-9732-761c3529d6a7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lazy-loading 'flavor' on Instance uuid 22cea165-cfde-403f-a0e2-82a86f5baa51 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1126.183902] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52184110-ada7-4bea-c2cd-c457009629d2, 'name': SearchDatastore_Task, 'duration_secs': 0.009467} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.184272] env[62753]: DEBUG oslo_concurrency.lockutils [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.184534] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1126.184774] env[62753]: DEBUG oslo_concurrency.lockutils [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1126.184927] env[62753]: DEBUG oslo_concurrency.lockutils [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1126.185122] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1126.185420] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c5331cfc-01f3-437b-b2f9-cb0de819d37f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.188794] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6382cd9d-7f3e-4f11-b862-3de07f913135 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.194290] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1126.194468] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1126.196719] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-513625f1-c238-4ee6-8463-e0bcb0c9e7c0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.199480] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f28c22c8-3841-48bd-873f-9fc0ed748529 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.205288] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1126.205288] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c49da7-6938-d760-ff86-c2b6f60887c6" [ 1126.205288] env[62753]: _type = "Task" [ 1126.205288] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.233216] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9f911744-39a3-4aa7-aa68-21d46fecb7da {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.238598] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c49da7-6938-d760-ff86-c2b6f60887c6, 'name': SearchDatastore_Task, 'duration_secs': 0.008937} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.239676] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-83f7c83b-c061-4758-81d5-5f59900d7234 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.244799] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-246bdba4-e8ff-4b13-a0fb-97cfa5a91649 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.249176] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1126.249176] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52fbbde1-7dd0-d06f-6c04-359e3954ded4" [ 1126.249176] env[62753]: _type = "Task" [ 1126.249176] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.259450] env[62753]: DEBUG nova.compute.provider_tree [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1126.265584] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52fbbde1-7dd0-d06f-6c04-359e3954ded4, 'name': SearchDatastore_Task, 'duration_secs': 0.008365} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.265818] env[62753]: DEBUG oslo_concurrency.lockutils [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1126.266080] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] ec0b07b8-1cb9-4911-8b03-efb466a21888/ec0b07b8-1cb9-4911-8b03-efb466a21888.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1126.266310] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-8fc99df0-0ae5-4b54-b7e1-4aa83d6afd1b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.273156] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1126.273156] env[62753]: value = "task-1332862" [ 1126.273156] env[62753]: _type = "Task" [ 1126.273156] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.280683] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332862, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.379124] env[62753]: DEBUG nova.objects.instance [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lazy-loading 'flavor' on Instance uuid ca01cd3f-7f0e-4c61-a704-74f9945792b4 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1126.401509] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-067cb155-e3fe-47a2-9732-761c3529d6a7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1126.401856] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1ce48f1e-9e2e-4abc-9684-e9a583db1555 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.411362] env[62753]: DEBUG oslo_vmware.api [None req-067cb155-e3fe-47a2-9732-761c3529d6a7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1126.411362] env[62753]: value = "task-1332863" [ 1126.411362] env[62753]: _type = "Task" [ 1126.411362] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.421100] env[62753]: DEBUG oslo_vmware.api [None req-067cb155-e3fe-47a2-9732-761c3529d6a7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332863, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.762765] env[62753]: DEBUG nova.scheduler.client.report [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1126.782999] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332862, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.485049} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.783287] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] ec0b07b8-1cb9-4911-8b03-efb466a21888/ec0b07b8-1cb9-4911-8b03-efb466a21888.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1126.783522] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1126.783772] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-1e6acf3c-e3de-44f2-a7ae-4cdd7c932c47 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1126.790739] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1126.790739] env[62753]: value = "task-1332864" [ 1126.790739] env[62753]: _type = "Task" [ 1126.790739] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1126.798872] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332864, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1126.884345] env[62753]: DEBUG oslo_concurrency.lockutils [None req-13e47cd6-f4f4-46c8-b860-8ff822df3166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.251s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1126.923369] env[62753]: DEBUG oslo_vmware.api [None req-067cb155-e3fe-47a2-9732-761c3529d6a7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332863, 'name': PowerOffVM_Task, 'duration_secs': 0.289517} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1126.923637] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-067cb155-e3fe-47a2-9732-761c3529d6a7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1126.923816] env[62753]: DEBUG nova.compute.manager [None req-067cb155-e3fe-47a2-9732-761c3529d6a7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1126.924613] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89fafccd-1c9f-4502-98a6-d6c798880779 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.267571] env[62753]: DEBUG oslo_concurrency.lockutils [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.169s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.268100] env[62753]: DEBUG nova.compute.manager [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1127.300605] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332864, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.062109} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.300889] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1127.302246] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8381583a-dc0c-40b2-b944-62650792aa49 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.324533] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] ec0b07b8-1cb9-4911-8b03-efb466a21888/ec0b07b8-1cb9-4911-8b03-efb466a21888.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1127.325188] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-34616238-cfc2-43c0-b6f3-87de22c9be31 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.345437] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1127.345437] env[62753]: value = "task-1332865" [ 1127.345437] env[62753]: _type = "Task" [ 1127.345437] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.355287] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332865, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1127.436021] env[62753]: DEBUG oslo_concurrency.lockutils [None req-067cb155-e3fe-47a2-9732-761c3529d6a7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.550s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1127.511859] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1127.512455] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1127.512703] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1127.512876] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1127.513031] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62753) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1127.757236] env[62753]: DEBUG nova.compute.manager [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Stashing vm_state: active {{(pid=62753) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1127.772541] env[62753]: DEBUG nova.compute.utils [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1127.773990] env[62753]: DEBUG nova.compute.manager [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1127.774198] env[62753]: DEBUG nova.network.neutron [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1127.818523] env[62753]: DEBUG nova.policy [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '592d1d366cc4461299dbc28cee63e5b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8925b84dcf9a47fbaf2eb8044b3850fa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1127.856589] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332865, 'name': ReconfigVM_Task, 'duration_secs': 0.267312} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1127.856771] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Reconfigured VM instance instance-0000006b to attach disk [datastore2] ec0b07b8-1cb9-4911-8b03-efb466a21888/ec0b07b8-1cb9-4911-8b03-efb466a21888.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1127.857432] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-b0af4b5c-faf7-476f-a9cc-86e0cef7007f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1127.867426] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1127.867426] env[62753]: value = "task-1332866" [ 1127.867426] env[62753]: _type = "Task" [ 1127.867426] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1127.881501] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332866, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.006878] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.129928] env[62753]: DEBUG nova.network.neutron [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Successfully created port: f7c88469-8383-42ae-a124-f4fdadc03fa2 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1128.278089] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1128.278591] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1128.281050] env[62753]: DEBUG nova.compute.manager [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1128.292684] env[62753]: DEBUG nova.objects.instance [None req-25c7fec7-7954-40b4-9331-9090be2f1ea9 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lazy-loading 'flavor' on Instance uuid 22cea165-cfde-403f-a0e2-82a86f5baa51 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1128.377652] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332866, 'name': Rename_Task, 'duration_secs': 0.164627} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.377958] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1128.378234] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-a0a65838-ac89-41ef-a571-24a5d886efbf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1128.386459] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1128.386459] env[62753]: value = "task-1332867" [ 1128.386459] env[62753]: _type = "Task" [ 1128.386459] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1128.396257] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332867, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1128.513234] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1128.789909] env[62753]: INFO nova.compute.claims [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1128.799243] env[62753]: DEBUG oslo_concurrency.lockutils [None req-25c7fec7-7954-40b4-9331-9090be2f1ea9 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1128.799243] env[62753]: DEBUG oslo_concurrency.lockutils [None req-25c7fec7-7954-40b4-9331-9090be2f1ea9 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1128.799426] env[62753]: DEBUG nova.network.neutron [None req-25c7fec7-7954-40b4-9331-9090be2f1ea9 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1128.799459] env[62753]: DEBUG nova.objects.instance [None req-25c7fec7-7954-40b4-9331-9090be2f1ea9 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lazy-loading 'info_cache' on Instance uuid 22cea165-cfde-403f-a0e2-82a86f5baa51 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1128.896756] env[62753]: DEBUG oslo_vmware.api [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332867, 'name': PowerOnVM_Task, 'duration_secs': 0.454368} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1128.897133] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1128.897368] env[62753]: INFO nova.compute.manager [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Took 6.80 seconds to spawn the instance on the hypervisor. [ 1128.897556] env[62753]: DEBUG nova.compute.manager [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1128.898350] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3de0569-dd3d-4842-9b5a-753709e2e770 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.212688] env[62753]: DEBUG oslo_concurrency.lockutils [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Acquiring lock "6a97171d-7ea2-4581-8707-b6be33f1e8ed" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.212947] env[62753]: DEBUG oslo_concurrency.lockutils [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Lock "6a97171d-7ea2-4581-8707-b6be33f1e8ed" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.298934] env[62753]: INFO nova.compute.resource_tracker [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating resource usage from migration ecbfc80d-d1c3-449f-b2a0-44dbf0e521de [ 1129.302588] env[62753]: DEBUG nova.compute.manager [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1129.305107] env[62753]: DEBUG nova.objects.base [None req-25c7fec7-7954-40b4-9331-9090be2f1ea9 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Object Instance<22cea165-cfde-403f-a0e2-82a86f5baa51> lazy-loaded attributes: flavor,info_cache {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1129.330322] env[62753]: DEBUG nova.virt.hardware [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1129.330586] env[62753]: DEBUG nova.virt.hardware [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1129.330751] env[62753]: DEBUG nova.virt.hardware [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1129.330935] env[62753]: DEBUG nova.virt.hardware [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1129.331100] env[62753]: DEBUG nova.virt.hardware [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1129.331256] env[62753]: DEBUG nova.virt.hardware [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1129.331469] env[62753]: DEBUG nova.virt.hardware [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1129.331638] env[62753]: DEBUG nova.virt.hardware [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1129.331813] env[62753]: DEBUG nova.virt.hardware [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1129.331979] env[62753]: DEBUG nova.virt.hardware [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1129.332170] env[62753]: DEBUG nova.virt.hardware [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1129.333289] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ee67b6a0-2750-4730-ac1f-b874e580382d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.343747] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-11e5d5e5-20c1-4dcf-95c9-e56fc483306e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.415143] env[62753]: INFO nova.compute.manager [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Took 11.52 seconds to build instance. [ 1129.423601] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5415c202-804f-4ecf-b106-70500cab3341 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.432284] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1acc70aa-30e9-4b42-b2c6-c0a82f456fa5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.462159] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2a110405-cd75-4d22-8abc-6e2a9ebf4894 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.470297] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-549bb9c1-13cd-48c9-92e0-2dcaaabed3c0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1129.484022] env[62753]: DEBUG nova.compute.provider_tree [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1129.702391] env[62753]: DEBUG nova.compute.manager [req-3d1c60c9-de62-43c4-b2a4-0b595f1a9056 req-fe6921da-a302-44c0-a013-aadf78802fa4 service nova] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Received event network-vif-plugged-f7c88469-8383-42ae-a124-f4fdadc03fa2 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1129.702595] env[62753]: DEBUG oslo_concurrency.lockutils [req-3d1c60c9-de62-43c4-b2a4-0b595f1a9056 req-fe6921da-a302-44c0-a013-aadf78802fa4 service nova] Acquiring lock "21e6cf03-e528-4289-8813-86020a4e8d2a-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1129.702805] env[62753]: DEBUG oslo_concurrency.lockutils [req-3d1c60c9-de62-43c4-b2a4-0b595f1a9056 req-fe6921da-a302-44c0-a013-aadf78802fa4 service nova] Lock "21e6cf03-e528-4289-8813-86020a4e8d2a-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1129.702976] env[62753]: DEBUG oslo_concurrency.lockutils [req-3d1c60c9-de62-43c4-b2a4-0b595f1a9056 req-fe6921da-a302-44c0-a013-aadf78802fa4 service nova] Lock "21e6cf03-e528-4289-8813-86020a4e8d2a-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.703169] env[62753]: DEBUG nova.compute.manager [req-3d1c60c9-de62-43c4-b2a4-0b595f1a9056 req-fe6921da-a302-44c0-a013-aadf78802fa4 service nova] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] No waiting events found dispatching network-vif-plugged-f7c88469-8383-42ae-a124-f4fdadc03fa2 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1129.703337] env[62753]: WARNING nova.compute.manager [req-3d1c60c9-de62-43c4-b2a4-0b595f1a9056 req-fe6921da-a302-44c0-a013-aadf78802fa4 service nova] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Received unexpected event network-vif-plugged-f7c88469-8383-42ae-a124-f4fdadc03fa2 for instance with vm_state building and task_state spawning. [ 1129.714682] env[62753]: DEBUG nova.compute.manager [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1129.787420] env[62753]: DEBUG nova.network.neutron [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Successfully updated port: f7c88469-8383-42ae-a124-f4fdadc03fa2 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1129.917363] env[62753]: DEBUG oslo_concurrency.lockutils [None req-95ac598a-6fc4-4994-acce-62f992edeca0 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.034s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1129.987630] env[62753]: DEBUG nova.scheduler.client.report [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1130.034437] env[62753]: DEBUG nova.network.neutron [None req-25c7fec7-7954-40b4-9331-9090be2f1ea9 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance_info_cache with network_info: [{"id": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "address": "fa:16:3e:88:6a:1d", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25e4c4c8-9d", "ovs_interfaceid": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.040051] env[62753]: DEBUG nova.compute.manager [req-6ae36ac1-5ba5-40d6-aa4b-92d6b8e65b8c req-9691491b-b2f5-458f-8f4a-d294cbcdd03e service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Received event network-changed-8600276a-3346-44f1-a760-df6caca2e51e {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1130.040051] env[62753]: DEBUG nova.compute.manager [req-6ae36ac1-5ba5-40d6-aa4b-92d6b8e65b8c req-9691491b-b2f5-458f-8f4a-d294cbcdd03e service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Refreshing instance network info cache due to event network-changed-8600276a-3346-44f1-a760-df6caca2e51e. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1130.040283] env[62753]: DEBUG oslo_concurrency.lockutils [req-6ae36ac1-5ba5-40d6-aa4b-92d6b8e65b8c req-9691491b-b2f5-458f-8f4a-d294cbcdd03e service nova] Acquiring lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.040437] env[62753]: DEBUG oslo_concurrency.lockutils [req-6ae36ac1-5ba5-40d6-aa4b-92d6b8e65b8c req-9691491b-b2f5-458f-8f4a-d294cbcdd03e service nova] Acquired lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.041614] env[62753]: DEBUG nova.network.neutron [req-6ae36ac1-5ba5-40d6-aa4b-92d6b8e65b8c req-9691491b-b2f5-458f-8f4a-d294cbcdd03e service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Refreshing network info cache for port 8600276a-3346-44f1-a760-df6caca2e51e {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1130.235664] env[62753]: DEBUG oslo_concurrency.lockutils [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1130.289920] env[62753]: DEBUG oslo_concurrency.lockutils [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "refresh_cache-21e6cf03-e528-4289-8813-86020a4e8d2a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1130.290076] env[62753]: DEBUG oslo_concurrency.lockutils [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquired lock "refresh_cache-21e6cf03-e528-4289-8813-86020a4e8d2a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1130.290244] env[62753]: DEBUG nova.network.neutron [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1130.493027] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.214s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1130.493027] env[62753]: INFO nova.compute.manager [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Migrating [ 1130.500582] env[62753]: DEBUG oslo_concurrency.lockutils [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.265s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1130.502147] env[62753]: INFO nova.compute.claims [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1130.537465] env[62753]: DEBUG oslo_concurrency.lockutils [None req-25c7fec7-7954-40b4-9331-9090be2f1ea9 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1130.785589] env[62753]: DEBUG nova.network.neutron [req-6ae36ac1-5ba5-40d6-aa4b-92d6b8e65b8c req-9691491b-b2f5-458f-8f4a-d294cbcdd03e service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Updated VIF entry in instance network info cache for port 8600276a-3346-44f1-a760-df6caca2e51e. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1130.786046] env[62753]: DEBUG nova.network.neutron [req-6ae36ac1-5ba5-40d6-aa4b-92d6b8e65b8c req-9691491b-b2f5-458f-8f4a-d294cbcdd03e service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Updating instance_info_cache with network_info: [{"id": "8600276a-3346-44f1-a760-df6caca2e51e", "address": "fa:16:3e:bf:ad:12", "network": {"id": "939bb8aa-f8d4-44cb-960f-c5a5fbcf99c8", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1149502111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dd81bdb01ff4c39a4959f4af2acf61a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8600276a-33", "ovs_interfaceid": "8600276a-3346-44f1-a760-df6caca2e51e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1130.823108] env[62753]: DEBUG nova.network.neutron [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1130.952887] env[62753]: DEBUG nova.network.neutron [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Updating instance_info_cache with network_info: [{"id": "f7c88469-8383-42ae-a124-f4fdadc03fa2", "address": "fa:16:3e:ed:31:57", "network": {"id": "61c6e640-d2e6-4b3b-b191-7da9249dd339", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-703836479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8925b84dcf9a47fbaf2eb8044b3850fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7c88469-83", "ovs_interfaceid": "f7c88469-8383-42ae-a124-f4fdadc03fa2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1131.010343] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1131.010343] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1131.017267] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.017267] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.017470] env[62753]: DEBUG nova.network.neutron [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1131.040938] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-25c7fec7-7954-40b4-9331-9090be2f1ea9 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1131.041245] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d6b9550e-83be-4543-864b-bfc615827f88 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.050471] env[62753]: DEBUG oslo_vmware.api [None req-25c7fec7-7954-40b4-9331-9090be2f1ea9 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1131.050471] env[62753]: value = "task-1332868" [ 1131.050471] env[62753]: _type = "Task" [ 1131.050471] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.059015] env[62753]: DEBUG oslo_vmware.api [None req-25c7fec7-7954-40b4-9331-9090be2f1ea9 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332868, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.288966] env[62753]: DEBUG oslo_concurrency.lockutils [req-6ae36ac1-5ba5-40d6-aa4b-92d6b8e65b8c req-9691491b-b2f5-458f-8f4a-d294cbcdd03e service nova] Releasing lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.456207] env[62753]: DEBUG oslo_concurrency.lockutils [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Releasing lock "refresh_cache-21e6cf03-e528-4289-8813-86020a4e8d2a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1131.456562] env[62753]: DEBUG nova.compute.manager [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Instance network_info: |[{"id": "f7c88469-8383-42ae-a124-f4fdadc03fa2", "address": "fa:16:3e:ed:31:57", "network": {"id": "61c6e640-d2e6-4b3b-b191-7da9249dd339", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-703836479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8925b84dcf9a47fbaf2eb8044b3850fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7c88469-83", "ovs_interfaceid": "f7c88469-8383-42ae-a124-f4fdadc03fa2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1131.457070] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:ed:31:57', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '510d3c47-3615-43d5-aa5d-a279fd915e71', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'f7c88469-8383-42ae-a124-f4fdadc03fa2', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1131.465617] env[62753]: DEBUG oslo.service.loopingcall [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1131.465866] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1131.466160] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-122d1ac6-b42e-4a4e-8c95-7599fb005124 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.488248] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1131.488248] env[62753]: value = "task-1332869" [ 1131.488248] env[62753]: _type = "Task" [ 1131.488248] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1131.496333] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332869, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1131.513573] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1131.564688] env[62753]: DEBUG oslo_vmware.api [None req-25c7fec7-7954-40b4-9331-9090be2f1ea9 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332868, 'name': PowerOnVM_Task, 'duration_secs': 0.379748} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1131.567224] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-25c7fec7-7954-40b4-9331-9090be2f1ea9 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1131.567432] env[62753]: DEBUG nova.compute.manager [None req-25c7fec7-7954-40b4-9331-9090be2f1ea9 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1131.568452] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2906133e-7010-43ce-a8e8-b7fcebfefc0e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.643015] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8aee8454-3589-469d-b217-e8f07bb621bc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.654533] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-35d7fb67-4329-4d76-bdbb-79077be34fec {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.688339] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8030ac77-363f-41f4-98b2-2c6c9d595ed5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.697244] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c19292a9-9db4-49b6-a00f-0e75e3e16f01 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1131.712159] env[62753]: DEBUG nova.compute.provider_tree [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1131.728325] env[62753]: DEBUG nova.compute.manager [req-de98cc38-b6c9-42aa-b0b5-d8e8564c418a req-591ea630-62b2-4598-92ab-08a9e714f51b service nova] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Received event network-changed-f7c88469-8383-42ae-a124-f4fdadc03fa2 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1131.728525] env[62753]: DEBUG nova.compute.manager [req-de98cc38-b6c9-42aa-b0b5-d8e8564c418a req-591ea630-62b2-4598-92ab-08a9e714f51b service nova] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Refreshing instance network info cache due to event network-changed-f7c88469-8383-42ae-a124-f4fdadc03fa2. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1131.728753] env[62753]: DEBUG oslo_concurrency.lockutils [req-de98cc38-b6c9-42aa-b0b5-d8e8564c418a req-591ea630-62b2-4598-92ab-08a9e714f51b service nova] Acquiring lock "refresh_cache-21e6cf03-e528-4289-8813-86020a4e8d2a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1131.728949] env[62753]: DEBUG oslo_concurrency.lockutils [req-de98cc38-b6c9-42aa-b0b5-d8e8564c418a req-591ea630-62b2-4598-92ab-08a9e714f51b service nova] Acquired lock "refresh_cache-21e6cf03-e528-4289-8813-86020a4e8d2a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1131.729168] env[62753]: DEBUG nova.network.neutron [req-de98cc38-b6c9-42aa-b0b5-d8e8564c418a req-591ea630-62b2-4598-92ab-08a9e714f51b service nova] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Refreshing network info cache for port f7c88469-8383-42ae-a124-f4fdadc03fa2 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1131.771794] env[62753]: DEBUG nova.network.neutron [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance_info_cache with network_info: [{"id": "ad95cf04-2066-4452-b160-65cf3cddbe87", "address": "fa:16:3e:eb:5e:01", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad95cf04-20", "ovs_interfaceid": "ad95cf04-2066-4452-b160-65cf3cddbe87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.000095] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332869, 'name': CreateVM_Task, 'duration_secs': 0.324409} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.000289] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1132.001016] env[62753]: DEBUG oslo_concurrency.lockutils [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.001189] env[62753]: DEBUG oslo_concurrency.lockutils [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.001532] env[62753]: DEBUG oslo_concurrency.lockutils [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1132.001806] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-13ec32ee-6100-4417-a5fe-b12d98801101 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.008183] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1132.008183] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5270235d-7c23-43e0-7b45-41476a8223dd" [ 1132.008183] env[62753]: _type = "Task" [ 1132.008183] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.018190] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5270235d-7c23-43e0-7b45-41476a8223dd, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.215334] env[62753]: DEBUG nova.scheduler.client.report [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1132.274482] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.430109] env[62753]: DEBUG nova.network.neutron [req-de98cc38-b6c9-42aa-b0b5-d8e8564c418a req-591ea630-62b2-4598-92ab-08a9e714f51b service nova] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Updated VIF entry in instance network info cache for port f7c88469-8383-42ae-a124-f4fdadc03fa2. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1132.430712] env[62753]: DEBUG nova.network.neutron [req-de98cc38-b6c9-42aa-b0b5-d8e8564c418a req-591ea630-62b2-4598-92ab-08a9e714f51b service nova] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Updating instance_info_cache with network_info: [{"id": "f7c88469-8383-42ae-a124-f4fdadc03fa2", "address": "fa:16:3e:ed:31:57", "network": {"id": "61c6e640-d2e6-4b3b-b191-7da9249dd339", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-703836479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8925b84dcf9a47fbaf2eb8044b3850fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7c88469-83", "ovs_interfaceid": "f7c88469-8383-42ae-a124-f4fdadc03fa2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1132.519256] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5270235d-7c23-43e0-7b45-41476a8223dd, 'name': SearchDatastore_Task, 'duration_secs': 0.012662} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1132.519604] env[62753]: DEBUG oslo_concurrency.lockutils [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1132.519855] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1132.520111] env[62753]: DEBUG oslo_concurrency.lockutils [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1132.520272] env[62753]: DEBUG oslo_concurrency.lockutils [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1132.520460] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1132.520729] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-34be932e-8d64-4e2e-8a8a-17213790ef34 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.536198] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1132.536437] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1132.537155] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3ef225eb-98db-4d73-8807-714c29126a70 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.542682] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1132.542682] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5209a2be-1917-aaab-668e-931fa52bf773" [ 1132.542682] env[62753]: _type = "Task" [ 1132.542682] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1132.550330] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5209a2be-1917-aaab-668e-931fa52bf773, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1132.720022] env[62753]: DEBUG oslo_concurrency.lockutils [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.219s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.720374] env[62753]: DEBUG nova.compute.manager [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1132.723049] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 1.210s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.723237] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1132.723408] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62753) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1132.724441] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-27eb7089-0feb-4689-ae55-23e682189644 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.733147] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c90b4b7-5987-4d8d-b2bf-fab9be026226 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.747289] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df2a52a7-7d71-4478-b878-a43a9b5d861f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.754240] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-082a999f-6896-41c4-89fa-13c2895ae689 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1132.788469] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180009MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=62753) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1132.788619] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1132.788807] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1132.933461] env[62753]: DEBUG oslo_concurrency.lockutils [req-de98cc38-b6c9-42aa-b0b5-d8e8564c418a req-591ea630-62b2-4598-92ab-08a9e714f51b service nova] Releasing lock "refresh_cache-21e6cf03-e528-4289-8813-86020a4e8d2a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.056138] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5209a2be-1917-aaab-668e-931fa52bf773, 'name': SearchDatastore_Task, 'duration_secs': 0.041127} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.056138] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-d77e16fd-f3a4-4168-b14a-085af487750d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.061964] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1133.061964] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c252e2-a274-237a-afd9-5c95af99aefb" [ 1133.061964] env[62753]: _type = "Task" [ 1133.061964] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.070685] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c252e2-a274-237a-afd9-5c95af99aefb, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.226168] env[62753]: DEBUG nova.compute.utils [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1133.227714] env[62753]: DEBUG nova.compute.manager [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1133.227910] env[62753]: DEBUG nova.network.neutron [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1133.268622] env[62753]: DEBUG nova.policy [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'baf880bf38f04cd7b61c931491a1f9e4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '36e8dd871c2444ee94b682acfca002aa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1133.535662] env[62753]: DEBUG nova.network.neutron [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Successfully created port: 0dd198e5-3992-4c1f-939a-16be88877273 {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1133.573664] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52c252e2-a274-237a-afd9-5c95af99aefb, 'name': SearchDatastore_Task, 'duration_secs': 0.00972} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1133.573941] env[62753]: DEBUG oslo_concurrency.lockutils [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1133.574442] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 21e6cf03-e528-4289-8813-86020a4e8d2a/21e6cf03-e528-4289-8813-86020a4e8d2a.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1133.574549] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-e163e6e4-7ce9-4fd2-b029-268e7adac98e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.582364] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1133.582364] env[62753]: value = "task-1332870" [ 1133.582364] env[62753]: _type = "Task" [ 1133.582364] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1133.590638] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332870, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1133.731830] env[62753]: DEBUG nova.compute.manager [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1133.805190] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Applying migration context for instance ca01cd3f-7f0e-4c61-a704-74f9945792b4 as it has an incoming, in-progress migration ecbfc80d-d1c3-449f-b2a0-44dbf0e521de. Migration status is migrating {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1133.806338] env[62753]: INFO nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating resource usage from migration ecbfc80d-d1c3-449f-b2a0-44dbf0e521de [ 1133.810802] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bed85393-095a-48de-926e-52b56c947bb9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1133.834697] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance 'ca01cd3f-7f0e-4c61-a704-74f9945792b4' progress to 0 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1133.965039] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 22cea165-cfde-403f-a0e2-82a86f5baa51 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1133.965244] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance a2e7fa98-566e-40f1-8da7-9318de89c14c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1133.965417] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 51790f51-c464-48cc-9891-41cae5a8f63e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1133.965615] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance ec0b07b8-1cb9-4911-8b03-efb466a21888 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1133.965744] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 21e6cf03-e528-4289-8813-86020a4e8d2a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1133.965864] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Migration ecbfc80d-d1c3-449f-b2a0-44dbf0e521de is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1133.965977] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance ca01cd3f-7f0e-4c61-a704-74f9945792b4 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1133.966110] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 6a97171d-7ea2-4581-8707-b6be33f1e8ed actively managed on this compute host and has allocations in placement: {'resources': {'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1133.966450] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Total usable vcpus: 48, total allocated vcpus: 8 {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1133.966638] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=2112MB phys_disk=200GB used_disk=7GB total_vcpus=48 used_vcpus=8 pci_stats=[] {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1133.987098] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Refreshing inventories for resource provider 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1134.001358] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Updating ProviderTree inventory for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1134.001580] env[62753]: DEBUG nova.compute.provider_tree [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Updating inventory in ProviderTree for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1134.014180] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Refreshing aggregate associations for resource provider 42c8603e-76bc-4e29-ad77-cac0094f7308, aggregates: None {{(pid=62753) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1134.033970] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Refreshing trait associations for resource provider 42c8603e-76bc-4e29-ad77-cac0094f7308, traits: COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62753) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1134.093279] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332870, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.483733} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.095665] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] 21e6cf03-e528-4289-8813-86020a4e8d2a/21e6cf03-e528-4289-8813-86020a4e8d2a.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1134.095889] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1134.097041] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d96b2d2a-11a3-45b8-815e-a08d83b1ea02 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.103626] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1134.103626] env[62753]: value = "task-1332871" [ 1134.103626] env[62753]: _type = "Task" [ 1134.103626] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.114571] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332871, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.154499] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5b3a9463-b168-4411-9adb-d0e2fbd41ffe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.162709] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e99c0f-a935-4f3b-830d-df4cb68954b9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.193081] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c74a02c-c79f-4a31-a41e-144de8787352 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.201043] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-87d11aab-e09e-4cd2-902f-7c4f1383074b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.213923] env[62753]: DEBUG nova.compute.provider_tree [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1134.236909] env[62753]: INFO nova.virt.block_device [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Booting with volume e055f16d-248a-4dca-97d1-ce25b21061b5 at /dev/sda [ 1134.270641] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-33882c1f-bc7b-4366-9c27-1bb223871576 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.280586] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56c04437-0595-4a3a-bd94-837e1b63440c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.309381] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-2f31ce90-fa78-4044-98c5-fb9af35250d7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.317854] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c7213714-21cd-481a-9c51-79a82d5455da {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.345101] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1134.345436] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9d16f3af-b31d-43b0-aea2-c864993ca4e0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.347311] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4db110ae-fec4-4525-826f-ef634e51b4aa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.353921] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a7685050-c075-4f19-94dc-9dd6b432d10b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.357578] env[62753]: DEBUG oslo_vmware.api [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1134.357578] env[62753]: value = "task-1332872" [ 1134.357578] env[62753]: _type = "Task" [ 1134.357578] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.370542] env[62753]: DEBUG oslo_vmware.api [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332872, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.374226] env[62753]: DEBUG nova.virt.block_device [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Updating existing volume attachment record: 0bfbece8-eb5a-40e3-a2ca-89eea719ea4b {{(pid=62753) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1134.613919] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332871, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.067818} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.614304] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1134.614947] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-349f51ae-e95e-432b-83cf-c0c56ed9b440 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.637356] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Reconfiguring VM instance instance-0000006c to attach disk [datastore2] 21e6cf03-e528-4289-8813-86020a4e8d2a/21e6cf03-e528-4289-8813-86020a4e8d2a.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1134.637632] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-2c653bbb-f7ca-4295-ad9b-cb87e10f1591 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1134.658127] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1134.658127] env[62753]: value = "task-1332873" [ 1134.658127] env[62753]: _type = "Task" [ 1134.658127] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1134.666780] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332873, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1134.717471] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1134.869438] env[62753]: DEBUG oslo_vmware.api [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332872, 'name': PowerOffVM_Task, 'duration_secs': 0.448762} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1134.869662] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1134.869851] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance 'ca01cd3f-7f0e-4c61-a704-74f9945792b4' progress to 17 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1134.932791] env[62753]: DEBUG nova.compute.manager [req-b7c13495-3c79-4f76-9c5f-177f4dbfdbb9 req-ade6ea34-e706-49e8-9ae3-2c42e39dd102 service nova] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Received event network-vif-plugged-0dd198e5-3992-4c1f-939a-16be88877273 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1134.933237] env[62753]: DEBUG oslo_concurrency.lockutils [req-b7c13495-3c79-4f76-9c5f-177f4dbfdbb9 req-ade6ea34-e706-49e8-9ae3-2c42e39dd102 service nova] Acquiring lock "6a97171d-7ea2-4581-8707-b6be33f1e8ed-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1134.933509] env[62753]: DEBUG oslo_concurrency.lockutils [req-b7c13495-3c79-4f76-9c5f-177f4dbfdbb9 req-ade6ea34-e706-49e8-9ae3-2c42e39dd102 service nova] Lock "6a97171d-7ea2-4581-8707-b6be33f1e8ed-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1134.933813] env[62753]: DEBUG oslo_concurrency.lockutils [req-b7c13495-3c79-4f76-9c5f-177f4dbfdbb9 req-ade6ea34-e706-49e8-9ae3-2c42e39dd102 service nova] Lock "6a97171d-7ea2-4581-8707-b6be33f1e8ed-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1134.934082] env[62753]: DEBUG nova.compute.manager [req-b7c13495-3c79-4f76-9c5f-177f4dbfdbb9 req-ade6ea34-e706-49e8-9ae3-2c42e39dd102 service nova] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] No waiting events found dispatching network-vif-plugged-0dd198e5-3992-4c1f-939a-16be88877273 {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1134.934333] env[62753]: WARNING nova.compute.manager [req-b7c13495-3c79-4f76-9c5f-177f4dbfdbb9 req-ade6ea34-e706-49e8-9ae3-2c42e39dd102 service nova] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Received unexpected event network-vif-plugged-0dd198e5-3992-4c1f-939a-16be88877273 for instance with vm_state building and task_state block_device_mapping. [ 1135.014220] env[62753]: DEBUG nova.network.neutron [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Successfully updated port: 0dd198e5-3992-4c1f-939a-16be88877273 {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1135.169324] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332873, 'name': ReconfigVM_Task, 'duration_secs': 0.270874} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.170123] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Reconfigured VM instance instance-0000006c to attach disk [datastore2] 21e6cf03-e528-4289-8813-86020a4e8d2a/21e6cf03-e528-4289-8813-86020a4e8d2a.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1135.170326] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-247c7caf-42bd-4bf7-a8b0-7acd0cc21bfc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.178514] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1135.178514] env[62753]: value = "task-1332874" [ 1135.178514] env[62753]: _type = "Task" [ 1135.178514] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.187424] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332874, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.223340] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1135.223568] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.435s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1135.223800] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1135.223942] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Cleaning up deleted instances {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1135.376217] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1135.376466] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1135.376663] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1135.376868] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1135.377033] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1135.377198] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1135.377408] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1135.377571] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1135.377744] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1135.377910] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1135.378100] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1135.383846] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-da702890-85a6-4f2e-b9ae-10889a5fb349 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.401242] env[62753]: DEBUG oslo_vmware.api [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1135.401242] env[62753]: value = "task-1332875" [ 1135.401242] env[62753]: _type = "Task" [ 1135.401242] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.410250] env[62753]: DEBUG oslo_vmware.api [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332875, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.519728] env[62753]: DEBUG oslo_concurrency.lockutils [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Acquiring lock "refresh_cache-6a97171d-7ea2-4581-8707-b6be33f1e8ed" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1135.519890] env[62753]: DEBUG oslo_concurrency.lockutils [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Acquired lock "refresh_cache-6a97171d-7ea2-4581-8707-b6be33f1e8ed" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1135.520062] env[62753]: DEBUG nova.network.neutron [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1135.689462] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332874, 'name': Rename_Task, 'duration_secs': 0.145095} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.689834] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1135.690015] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-e147cf9e-0951-476d-9577-c1b598a62e29 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1135.697821] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1135.697821] env[62753]: value = "task-1332876" [ 1135.697821] env[62753]: _type = "Task" [ 1135.697821] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1135.712740] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332876, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1135.736787] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] There are 46 instances to clean {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1135.737172] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 14f8f39a-3c29-4ba1-8c3f-10eeb7a236ac] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1135.911488] env[62753]: DEBUG oslo_vmware.api [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332875, 'name': ReconfigVM_Task, 'duration_secs': 0.448578} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1135.912418] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance 'ca01cd3f-7f0e-4c61-a704-74f9945792b4' progress to 33 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1136.056851] env[62753]: DEBUG nova.network.neutron [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1136.204263] env[62753]: DEBUG nova.network.neutron [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Updating instance_info_cache with network_info: [{"id": "0dd198e5-3992-4c1f-939a-16be88877273", "address": "fa:16:3e:8d:c6:d9", "network": {"id": "7b4c246f-28b0-49c5-8045-13fd80ec45d4", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-146238905-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "36e8dd871c2444ee94b682acfca002aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4813d311-2016-4f6a-a4b8-7613ab624fba", "external-id": "nsx-vlan-transportzone-386", "segmentation_id": 386, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0dd198e5-39", "ovs_interfaceid": "0dd198e5-3992-4c1f-939a-16be88877273", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1136.210493] env[62753]: DEBUG oslo_vmware.api [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332876, 'name': PowerOnVM_Task, 'duration_secs': 0.512512} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.210777] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1136.211043] env[62753]: INFO nova.compute.manager [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Took 6.91 seconds to spawn the instance on the hypervisor. [ 1136.211239] env[62753]: DEBUG nova.compute.manager [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1136.212122] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cf28abeb-6c82-481d-a657-ffe3e559ca29 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.240646] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 32563e1f-398f-4fc9-866f-d85d9f3f27a7] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1136.418784] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1136.419064] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1136.419231] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1136.419429] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1136.419572] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1136.419724] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1136.419935] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1136.420124] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1136.420295] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1136.420470] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1136.420675] env[62753]: DEBUG nova.virt.hardware [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1136.425968] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Reconfiguring VM instance instance-00000067 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1136.426281] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-a7d5d66b-7f6a-4a18-9268-d83df51fb305 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.446529] env[62753]: DEBUG oslo_vmware.api [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1136.446529] env[62753]: value = "task-1332877" [ 1136.446529] env[62753]: _type = "Task" [ 1136.446529] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.455290] env[62753]: DEBUG oslo_vmware.api [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332877, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.461639] env[62753]: DEBUG nova.compute.manager [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1136.462160] env[62753]: DEBUG nova.virt.hardware [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format=,created_at=,direct_url=,disk_format=,id=,min_disk=0,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=1073741824,status='active',tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1136.462388] env[62753]: DEBUG nova.virt.hardware [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1136.462551] env[62753]: DEBUG nova.virt.hardware [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1136.463134] env[62753]: DEBUG nova.virt.hardware [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1136.463326] env[62753]: DEBUG nova.virt.hardware [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1136.463531] env[62753]: DEBUG nova.virt.hardware [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1136.463785] env[62753]: DEBUG nova.virt.hardware [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1136.463965] env[62753]: DEBUG nova.virt.hardware [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1136.464160] env[62753]: DEBUG nova.virt.hardware [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1136.464329] env[62753]: DEBUG nova.virt.hardware [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1136.464509] env[62753]: DEBUG nova.virt.hardware [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1136.465695] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-870578f3-1a5d-4626-b853-5cb9595da3e5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.474395] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7f213ec8-94ac-4add-b529-9f0594a5affc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.707587] env[62753]: DEBUG oslo_concurrency.lockutils [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Releasing lock "refresh_cache-6a97171d-7ea2-4581-8707-b6be33f1e8ed" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1136.707936] env[62753]: DEBUG nova.compute.manager [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Instance network_info: |[{"id": "0dd198e5-3992-4c1f-939a-16be88877273", "address": "fa:16:3e:8d:c6:d9", "network": {"id": "7b4c246f-28b0-49c5-8045-13fd80ec45d4", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-146238905-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "36e8dd871c2444ee94b682acfca002aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4813d311-2016-4f6a-a4b8-7613ab624fba", "external-id": "nsx-vlan-transportzone-386", "segmentation_id": 386, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0dd198e5-39", "ovs_interfaceid": "0dd198e5-3992-4c1f-939a-16be88877273", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1136.708425] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:8d:c6:d9', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '4813d311-2016-4f6a-a4b8-7613ab624fba', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '0dd198e5-3992-4c1f-939a-16be88877273', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1136.716376] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Creating folder: Project (36e8dd871c2444ee94b682acfca002aa). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1136.716692] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1e46f7f9-da84-45a2-858b-762ce4b07261 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.731183] env[62753]: WARNING suds.client [-] Web service reported a SOAP processing fault using an unexpected HTTP status code 200. Reporting as an internal server error. [ 1136.731340] env[62753]: DEBUG oslo_vmware.api [-] Fault list: [DuplicateName] {{(pid=62753) _invoke_api /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:337}} [ 1136.731746] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Folder already exists: Project (36e8dd871c2444ee94b682acfca002aa). Parent ref: group-v284541. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1609}} [ 1136.731874] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Creating folder: Instances. Parent ref: group-v284689. {{(pid=62753) create_folder /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1599}} [ 1136.732118] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateFolder with opID=oslo.vmware-1689edcc-68cb-4dbb-a49f-b6f9a869294b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.734410] env[62753]: INFO nova.compute.manager [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Took 11.66 seconds to build instance. [ 1136.743740] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 6e417227-895d-4576-b025-7d4a0aafa379] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1136.746837] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Created folder: Instances in parent group-v284689. [ 1136.747121] env[62753]: DEBUG oslo.service.loopingcall [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1136.747371] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1136.747630] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-41cb06bb-7bc2-4ae5-aff3-1630695c36b1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.769043] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1136.769043] env[62753]: value = "task-1332880" [ 1136.769043] env[62753]: _type = "Task" [ 1136.769043] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1136.778746] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332880, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1136.957423] env[62753]: DEBUG oslo_vmware.api [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332877, 'name': ReconfigVM_Task, 'duration_secs': 0.23344} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1136.958856] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Reconfigured VM instance instance-00000067 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1136.960144] env[62753]: DEBUG nova.compute.manager [req-3c89c019-58ce-4b8b-9147-85c9e5b8fe82 req-10b5ca64-dffd-436e-a34c-258feea41403 service nova] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Received event network-changed-0dd198e5-3992-4c1f-939a-16be88877273 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1136.960338] env[62753]: DEBUG nova.compute.manager [req-3c89c019-58ce-4b8b-9147-85c9e5b8fe82 req-10b5ca64-dffd-436e-a34c-258feea41403 service nova] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Refreshing instance network info cache due to event network-changed-0dd198e5-3992-4c1f-939a-16be88877273. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1136.960560] env[62753]: DEBUG oslo_concurrency.lockutils [req-3c89c019-58ce-4b8b-9147-85c9e5b8fe82 req-10b5ca64-dffd-436e-a34c-258feea41403 service nova] Acquiring lock "refresh_cache-6a97171d-7ea2-4581-8707-b6be33f1e8ed" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1136.960710] env[62753]: DEBUG oslo_concurrency.lockutils [req-3c89c019-58ce-4b8b-9147-85c9e5b8fe82 req-10b5ca64-dffd-436e-a34c-258feea41403 service nova] Acquired lock "refresh_cache-6a97171d-7ea2-4581-8707-b6be33f1e8ed" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1136.960873] env[62753]: DEBUG nova.network.neutron [req-3c89c019-58ce-4b8b-9147-85c9e5b8fe82 req-10b5ca64-dffd-436e-a34c-258feea41403 service nova] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Refreshing network info cache for port 0dd198e5-3992-4c1f-939a-16be88877273 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1136.965944] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f0fa061-cb65-4be0-a372-6a0c891cc630 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1136.989985] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] ca01cd3f-7f0e-4c61-a704-74f9945792b4/ca01cd3f-7f0e-4c61-a704-74f9945792b4.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1136.991433] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e29687b6-82ed-4510-bf30-c99ded8dc9b6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.012329] env[62753]: DEBUG oslo_vmware.api [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1137.012329] env[62753]: value = "task-1332881" [ 1137.012329] env[62753]: _type = "Task" [ 1137.012329] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.022376] env[62753]: DEBUG oslo_vmware.api [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332881, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.237270] env[62753]: DEBUG oslo_concurrency.lockutils [None req-141f2b84-cd30-4b28-9260-03f84cbd0060 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "21e6cf03-e528-4289-8813-86020a4e8d2a" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 13.168s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1137.247943] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 01d9da0b-f4e5-474c-aab2-a0f4f0c99994] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1137.261186] env[62753]: DEBUG nova.compute.manager [req-2a850391-32a8-4b52-9a7e-3d0b2c0c0b8a req-f4fe55b5-41c7-4ada-a110-b7a563204432 service nova] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Received event network-changed-f7c88469-8383-42ae-a124-f4fdadc03fa2 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1137.261400] env[62753]: DEBUG nova.compute.manager [req-2a850391-32a8-4b52-9a7e-3d0b2c0c0b8a req-f4fe55b5-41c7-4ada-a110-b7a563204432 service nova] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Refreshing instance network info cache due to event network-changed-f7c88469-8383-42ae-a124-f4fdadc03fa2. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1137.261635] env[62753]: DEBUG oslo_concurrency.lockutils [req-2a850391-32a8-4b52-9a7e-3d0b2c0c0b8a req-f4fe55b5-41c7-4ada-a110-b7a563204432 service nova] Acquiring lock "refresh_cache-21e6cf03-e528-4289-8813-86020a4e8d2a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1137.261760] env[62753]: DEBUG oslo_concurrency.lockutils [req-2a850391-32a8-4b52-9a7e-3d0b2c0c0b8a req-f4fe55b5-41c7-4ada-a110-b7a563204432 service nova] Acquired lock "refresh_cache-21e6cf03-e528-4289-8813-86020a4e8d2a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1137.261925] env[62753]: DEBUG nova.network.neutron [req-2a850391-32a8-4b52-9a7e-3d0b2c0c0b8a req-f4fe55b5-41c7-4ada-a110-b7a563204432 service nova] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Refreshing network info cache for port f7c88469-8383-42ae-a124-f4fdadc03fa2 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1137.281239] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332880, 'name': CreateVM_Task, 'duration_secs': 0.329909} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1137.281376] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1137.282301] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Block device information present: {'root_device_name': '/dev/sda', 'image': [], 'ephemerals': [], 'block_device_mapping': [{'guest_format': None, 'delete_on_termination': True, 'attachment_id': '0bfbece8-eb5a-40e3-a2ca-89eea719ea4b', 'device_type': None, 'mount_device': '/dev/sda', 'disk_bus': None, 'boot_index': 0, 'connection_info': {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284692', 'volume_id': 'e055f16d-248a-4dca-97d1-ce25b21061b5', 'name': 'volume-e055f16d-248a-4dca-97d1-ce25b21061b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6a97171d-7ea2-4581-8707-b6be33f1e8ed', 'attached_at': '', 'detached_at': '', 'volume_id': 'e055f16d-248a-4dca-97d1-ce25b21061b5', 'serial': 'e055f16d-248a-4dca-97d1-ce25b21061b5'}, 'volume_type': None}], 'swap': None} {{(pid=62753) spawn /opt/stack/nova/nova/virt/vmwareapi/vmops.py:799}} [ 1137.282428] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Root volume attach. Driver type: vmdk {{(pid=62753) attach_root_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:661}} [ 1137.285099] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18c03684-8fad-4a3f-8bae-04c277663a1b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.293372] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6852efc4-33ed-4a1e-b28d-35c1a1b9d7e2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.301709] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-721478d7-4f3c-4813-927f-d85b77ff6c6b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.309229] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.RelocateVM_Task with opID=oslo.vmware-3fe6c1f8-2aab-4ebb-8a61-c0dbd78752e2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1137.319340] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Waiting for the task: (returnval){ [ 1137.319340] env[62753]: value = "task-1332882" [ 1137.319340] env[62753]: _type = "Task" [ 1137.319340] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1137.328915] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332882, 'name': RelocateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.523206] env[62753]: DEBUG oslo_vmware.api [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332881, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.675893] env[62753]: DEBUG nova.network.neutron [req-3c89c019-58ce-4b8b-9147-85c9e5b8fe82 req-10b5ca64-dffd-436e-a34c-258feea41403 service nova] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Updated VIF entry in instance network info cache for port 0dd198e5-3992-4c1f-939a-16be88877273. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1137.676290] env[62753]: DEBUG nova.network.neutron [req-3c89c019-58ce-4b8b-9147-85c9e5b8fe82 req-10b5ca64-dffd-436e-a34c-258feea41403 service nova] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Updating instance_info_cache with network_info: [{"id": "0dd198e5-3992-4c1f-939a-16be88877273", "address": "fa:16:3e:8d:c6:d9", "network": {"id": "7b4c246f-28b0-49c5-8045-13fd80ec45d4", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-146238905-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "36e8dd871c2444ee94b682acfca002aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4813d311-2016-4f6a-a4b8-7613ab624fba", "external-id": "nsx-vlan-transportzone-386", "segmentation_id": 386, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0dd198e5-39", "ovs_interfaceid": "0dd198e5-3992-4c1f-939a-16be88877273", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1137.751526] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: eb86a978-27fa-41b2-a7e5-4b3a8ba0a152] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1137.832036] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332882, 'name': RelocateVM_Task} progress is 42%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1137.998695] env[62753]: DEBUG nova.network.neutron [req-2a850391-32a8-4b52-9a7e-3d0b2c0c0b8a req-f4fe55b5-41c7-4ada-a110-b7a563204432 service nova] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Updated VIF entry in instance network info cache for port f7c88469-8383-42ae-a124-f4fdadc03fa2. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1137.999203] env[62753]: DEBUG nova.network.neutron [req-2a850391-32a8-4b52-9a7e-3d0b2c0c0b8a req-f4fe55b5-41c7-4ada-a110-b7a563204432 service nova] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Updating instance_info_cache with network_info: [{"id": "f7c88469-8383-42ae-a124-f4fdadc03fa2", "address": "fa:16:3e:ed:31:57", "network": {"id": "61c6e640-d2e6-4b3b-b191-7da9249dd339", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-703836479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.14", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.166", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8925b84dcf9a47fbaf2eb8044b3850fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapf7c88469-83", "ovs_interfaceid": "f7c88469-8383-42ae-a124-f4fdadc03fa2", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1138.026535] env[62753]: DEBUG oslo_vmware.api [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332881, 'name': ReconfigVM_Task, 'duration_secs': 0.587566} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1138.026886] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Reconfigured VM instance instance-00000067 to attach disk [datastore1] ca01cd3f-7f0e-4c61-a704-74f9945792b4/ca01cd3f-7f0e-4c61-a704-74f9945792b4.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1138.027239] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance 'ca01cd3f-7f0e-4c61-a704-74f9945792b4' progress to 50 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1138.179232] env[62753]: DEBUG oslo_concurrency.lockutils [req-3c89c019-58ce-4b8b-9147-85c9e5b8fe82 req-10b5ca64-dffd-436e-a34c-258feea41403 service nova] Releasing lock "refresh_cache-6a97171d-7ea2-4581-8707-b6be33f1e8ed" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1138.255767] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 0f81b85b-0550-46b9-b77c-9075a99cfaec] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1138.335552] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332882, 'name': RelocateVM_Task} progress is 54%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1138.502192] env[62753]: DEBUG oslo_concurrency.lockutils [req-2a850391-32a8-4b52-9a7e-3d0b2c0c0b8a req-f4fe55b5-41c7-4ada-a110-b7a563204432 service nova] Releasing lock "refresh_cache-21e6cf03-e528-4289-8813-86020a4e8d2a" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1138.535232] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bce4ae40-ca2a-4f7c-bd11-4b1b4b2c2168 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.562034] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-adc7aedb-8589-40d8-a7e2-e8226b6a2ca0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1138.587772] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance 'ca01cd3f-7f0e-4c61-a704-74f9945792b4' progress to 67 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1138.758988] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 961ced74-10af-4443-857e-545cd890674e] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1138.834869] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332882, 'name': RelocateVM_Task} progress is 69%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.262903] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: d7ba8be1-befb-4a0b-9c35-d86e7d586203] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1139.335455] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332882, 'name': RelocateVM_Task} progress is 82%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1139.767822] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: ff50c5f4-ab27-4f17-948e-80e2dcc623eb] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1139.833476] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332882, 'name': RelocateVM_Task} progress is 97%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.246239] env[62753]: DEBUG nova.network.neutron [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Port ad95cf04-2066-4452-b160-65cf3cddbe87 binding to destination host cpu-1 is already ACTIVE {{(pid=62753) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1140.270976] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: b1e4c8f9-471d-4cca-9a12-56191767a5b1] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1140.333189] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332882, 'name': RelocateVM_Task} progress is 97%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1140.774261] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: b5877d39-2f94-4a01-9db1-89a3e612095c] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1140.833857] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332882, 'name': RelocateVM_Task} progress is 97%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.267802] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1141.268095] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1141.268252] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1141.277523] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 9d7066e1-3721-42d2-8e80-91eacbbd8b80] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1141.333655] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332882, 'name': RelocateVM_Task, 'duration_secs': 3.907745} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.334152] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Volume attach. Driver type: vmdk {{(pid=62753) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1141.334364] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284692', 'volume_id': 'e055f16d-248a-4dca-97d1-ce25b21061b5', 'name': 'volume-e055f16d-248a-4dca-97d1-ce25b21061b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6a97171d-7ea2-4581-8707-b6be33f1e8ed', 'attached_at': '', 'detached_at': '', 'volume_id': 'e055f16d-248a-4dca-97d1-ce25b21061b5', 'serial': 'e055f16d-248a-4dca-97d1-ce25b21061b5'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1141.335108] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7bf9562c-e0b0-4bbf-a632-bc4b61d3607e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.350374] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-14ef7a59-0cd1-47b6-93c6-dbc69a41335a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.371320] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Reconfiguring VM instance instance-0000006d to attach disk [datastore1] volume-e055f16d-248a-4dca-97d1-ce25b21061b5/volume-e055f16d-248a-4dca-97d1-ce25b21061b5.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1141.371798] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f1a77307-aa6d-4935-820f-acd2ca03afec {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.392143] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Waiting for the task: (returnval){ [ 1141.392143] env[62753]: value = "task-1332883" [ 1141.392143] env[62753]: _type = "Task" [ 1141.392143] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.399835] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332883, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1141.780400] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 3009b1e4-53b1-4b12-8aaf-3c36b37c5d99] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1141.902141] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332883, 'name': ReconfigVM_Task, 'duration_secs': 0.28656} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1141.902452] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Reconfigured VM instance instance-0000006d to attach disk [datastore1] volume-e055f16d-248a-4dca-97d1-ce25b21061b5/volume-e055f16d-248a-4dca-97d1-ce25b21061b5.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1141.907247] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-002bd224-a2ae-41c8-8770-4626192d103a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1141.922705] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Waiting for the task: (returnval){ [ 1141.922705] env[62753]: value = "task-1332884" [ 1141.922705] env[62753]: _type = "Task" [ 1141.922705] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1141.938250] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332884, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.283299] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: e1c59139-ff02-453e-b54f-a56b1e40fdb5] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1142.303656] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1142.303870] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1142.304056] env[62753]: DEBUG nova.network.neutron [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1142.432066] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332884, 'name': ReconfigVM_Task, 'duration_secs': 0.1269} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.432375] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284692', 'volume_id': 'e055f16d-248a-4dca-97d1-ce25b21061b5', 'name': 'volume-e055f16d-248a-4dca-97d1-ce25b21061b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6a97171d-7ea2-4581-8707-b6be33f1e8ed', 'attached_at': '', 'detached_at': '', 'volume_id': 'e055f16d-248a-4dca-97d1-ce25b21061b5', 'serial': 'e055f16d-248a-4dca-97d1-ce25b21061b5'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1142.432902] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-7f5450f3-3935-4104-b669-b708966abbb2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.439061] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Waiting for the task: (returnval){ [ 1142.439061] env[62753]: value = "task-1332885" [ 1142.439061] env[62753]: _type = "Task" [ 1142.439061] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.446168] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332885, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1142.786382] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 04fd0022-f20a-4217-9e47-5381635b17a6] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1142.948677] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332885, 'name': Rename_Task, 'duration_secs': 0.132052} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1142.949012] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1142.949269] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-ee76c9d1-16cf-4bcd-880a-94547ce01378 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1142.955208] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Waiting for the task: (returnval){ [ 1142.955208] env[62753]: value = "task-1332886" [ 1142.955208] env[62753]: _type = "Task" [ 1142.955208] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1142.962486] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332886, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1143.017479] env[62753]: DEBUG nova.network.neutron [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance_info_cache with network_info: [{"id": "ad95cf04-2066-4452-b160-65cf3cddbe87", "address": "fa:16:3e:eb:5e:01", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad95cf04-20", "ovs_interfaceid": "ad95cf04-2066-4452-b160-65cf3cddbe87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1143.289148] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 91fc519e-c9bb-4605-bdc6-4f463561b988] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1143.465695] env[62753]: DEBUG oslo_vmware.api [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332886, 'name': PowerOnVM_Task, 'duration_secs': 0.41892} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1143.465969] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1143.466213] env[62753]: INFO nova.compute.manager [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Took 7.00 seconds to spawn the instance on the hypervisor. [ 1143.466399] env[62753]: DEBUG nova.compute.manager [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1143.467271] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aff35ed9-6bde-416d-b989-8f81cac9a713 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1143.520740] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1143.793204] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 2b90824a-5d49-4ec9-bcf6-4f868ca876cf] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1143.983623] env[62753]: INFO nova.compute.manager [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Took 13.76 seconds to build instance. [ 1144.031051] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e1485071-9f88-4ffa-87e0-61d674795fc5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.041114] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9b6cf3c8-ddb6-4f5b-9520-b5fa2bb9c3d5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1144.296200] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: bba5e3cb-c896-4ce7-ac7f-1cf2aba20465] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1144.486046] env[62753]: DEBUG oslo_concurrency.lockutils [None req-55afcc45-d649-414d-8ca8-4dd04bf691b5 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Lock "6a97171d-7ea2-4581-8707-b6be33f1e8ed" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 15.273s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1144.572029] env[62753]: DEBUG nova.compute.manager [req-455ff467-6a48-4c02-a44d-083fd3afc146 req-45a246b8-44ac-4451-93e4-cea476774fa4 service nova] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Received event network-changed-0dd198e5-3992-4c1f-939a-16be88877273 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1144.572029] env[62753]: DEBUG nova.compute.manager [req-455ff467-6a48-4c02-a44d-083fd3afc146 req-45a246b8-44ac-4451-93e4-cea476774fa4 service nova] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Refreshing instance network info cache due to event network-changed-0dd198e5-3992-4c1f-939a-16be88877273. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1144.572029] env[62753]: DEBUG oslo_concurrency.lockutils [req-455ff467-6a48-4c02-a44d-083fd3afc146 req-45a246b8-44ac-4451-93e4-cea476774fa4 service nova] Acquiring lock "refresh_cache-6a97171d-7ea2-4581-8707-b6be33f1e8ed" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1144.572206] env[62753]: DEBUG oslo_concurrency.lockutils [req-455ff467-6a48-4c02-a44d-083fd3afc146 req-45a246b8-44ac-4451-93e4-cea476774fa4 service nova] Acquired lock "refresh_cache-6a97171d-7ea2-4581-8707-b6be33f1e8ed" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1144.572268] env[62753]: DEBUG nova.network.neutron [req-455ff467-6a48-4c02-a44d-083fd3afc146 req-45a246b8-44ac-4451-93e4-cea476774fa4 service nova] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Refreshing network info cache for port 0dd198e5-3992-4c1f-939a-16be88877273 {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1144.799463] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 53a76eb0-8b0c-4e47-8387-75df4c0b191f] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1145.145939] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b612bc4-1464-4253-b894-045e327a0885 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.171201] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef0a17d6-6629-48fd-8b14-1cd7110ffeab {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.179925] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance 'ca01cd3f-7f0e-4c61-a704-74f9945792b4' progress to 83 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1145.304367] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: ff4b295e-2eb2-42dd-b353-08df755f04ea] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1145.332030] env[62753]: DEBUG nova.network.neutron [req-455ff467-6a48-4c02-a44d-083fd3afc146 req-45a246b8-44ac-4451-93e4-cea476774fa4 service nova] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Updated VIF entry in instance network info cache for port 0dd198e5-3992-4c1f-939a-16be88877273. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1145.332554] env[62753]: DEBUG nova.network.neutron [req-455ff467-6a48-4c02-a44d-083fd3afc146 req-45a246b8-44ac-4451-93e4-cea476774fa4 service nova] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Updating instance_info_cache with network_info: [{"id": "0dd198e5-3992-4c1f-939a-16be88877273", "address": "fa:16:3e:8d:c6:d9", "network": {"id": "7b4c246f-28b0-49c5-8045-13fd80ec45d4", "bridge": "br-int", "label": "tempest-ServerActionsV293TestJSON-146238905-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.6", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.134", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "36e8dd871c2444ee94b682acfca002aa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "4813d311-2016-4f6a-a4b8-7613ab624fba", "external-id": "nsx-vlan-transportzone-386", "segmentation_id": 386, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap0dd198e5-39", "ovs_interfaceid": "0dd198e5-3992-4c1f-939a-16be88877273", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1145.687645] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1145.688048] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-de136b58-5a18-49cc-8eb8-a578a857dd7d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1145.695485] env[62753]: DEBUG oslo_vmware.api [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1145.695485] env[62753]: value = "task-1332887" [ 1145.695485] env[62753]: _type = "Task" [ 1145.695485] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1145.709119] env[62753]: DEBUG oslo_vmware.api [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332887, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1145.807950] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: abd5ec6e-7b12-433b-9009-f9108ab4e318] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1145.836118] env[62753]: DEBUG oslo_concurrency.lockutils [req-455ff467-6a48-4c02-a44d-083fd3afc146 req-45a246b8-44ac-4451-93e4-cea476774fa4 service nova] Releasing lock "refresh_cache-6a97171d-7ea2-4581-8707-b6be33f1e8ed" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1146.209605] env[62753]: DEBUG oslo_vmware.api [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332887, 'name': PowerOnVM_Task, 'duration_secs': 0.460615} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1146.210247] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1146.210247] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4a2b45ce-27ff-4825-892d-af86a8721888 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance 'ca01cd3f-7f0e-4c61-a704-74f9945792b4' progress to 100 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1146.311517] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: c3679c66-2763-4a04-8d14-e103bd952798] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1146.815197] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 8066aeb2-7be4-46e3-aba6-124fef342a49] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1147.318832] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: f8470f38-0150-41c6-a3ee-aabb5e68d98b] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1147.822138] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: b45bfcb4-b2cd-468c-9595-62c155d520fb] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1148.324742] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: a401909c-5978-492b-a4fa-b17d38a5af44] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1148.828580] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: b5674964-1928-4ecb-b1fd-8f60a94b4270] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1149.033074] env[62753]: DEBUG nova.network.neutron [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Port ad95cf04-2066-4452-b160-65cf3cddbe87 binding to destination host cpu-1 is already ACTIVE {{(pid=62753) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1149.033389] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1149.033548] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1149.033723] env[62753]: DEBUG nova.network.neutron [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1149.332269] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: d669ac7d-0cf7-4c2d-b63c-cb6c64e6c075] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1149.743897] env[62753]: DEBUG nova.network.neutron [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance_info_cache with network_info: [{"id": "ad95cf04-2066-4452-b160-65cf3cddbe87", "address": "fa:16:3e:eb:5e:01", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad95cf04-20", "ovs_interfaceid": "ad95cf04-2066-4452-b160-65cf3cddbe87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1149.835918] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 77cdd901-cc96-4a6f-a696-a65f54a96b1a] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1150.246270] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1150.339225] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: c73b1ae0-dc98-47f7-babf-e96169384785] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1150.750903] env[62753]: DEBUG nova.compute.manager [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62753) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1150.842655] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 6b9a08b0-3cb3-474a-8193-953e74252a53] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1151.345695] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 9a8a148c-057d-4c91-8b7c-4e05e6ac0f66] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1151.850018] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: f03739f4-921f-46e5-b0e7-e9a99d96d621] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1151.852077] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1151.852336] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1152.355085] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 57eef548-a0f1-4f0b-a026-885a10d005c2] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1152.357190] env[62753]: DEBUG nova.objects.instance [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lazy-loading 'migration_context' on Instance uuid ca01cd3f-7f0e-4c61-a704-74f9945792b4 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1152.859717] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 33c1de85-aade-4c69-8fdb-7672d8c53300] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1152.974153] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88106d7c-60e0-4ff3-931f-d0db0c6887a5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1152.982436] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-71e1f117-b6b5-4bb2-b9ef-795eaa76e454 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.013803] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-140ebe8d-24d0-4b5c-aa12-75466720cd42 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.020396] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-66ff9ea8-7eee-45e0-92aa-f91fb2ea9d66 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1153.033219] env[62753]: DEBUG nova.compute.provider_tree [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1153.366049] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 50e1c1a1-7b8f-49cd-932d-03c920209634] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1153.536871] env[62753]: DEBUG nova.scheduler.client.report [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1153.869121] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 4d20e083-2959-453a-8875-47955bc02613] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1154.372100] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: c5dae9a3-052b-4f4c-ac79-84aeaded457a] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1154.547036] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.694s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1154.876724] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 49311ef5-733a-4146-87ac-876a121a8d4d] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1155.379540] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 4fdf8765-5e49-4ba6-ab3a-70a71f36fcb5] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1155.882712] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: a6ee010c-6305-4009-80e0-92a2c58bcd7b] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1156.083551] env[62753]: INFO nova.compute.manager [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Swapping old allocation on dict_keys(['42c8603e-76bc-4e29-ad77-cac0094f7308']) held by migration ecbfc80d-d1c3-449f-b2a0-44dbf0e521de for instance [ 1156.106917] env[62753]: DEBUG nova.scheduler.client.report [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Overwriting current allocation {'allocations': {'42c8603e-76bc-4e29-ad77-cac0094f7308': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 133}}, 'project_id': '196457eaa51643c38b4ef4fc6fbbcba3', 'user_id': 'af45e909b0c84662a8cfdf92896e5fe5', 'consumer_generation': 1} on consumer ca01cd3f-7f0e-4c61-a704-74f9945792b4 {{(pid=62753) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1156.179674] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1156.179886] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1156.180156] env[62753]: DEBUG nova.network.neutron [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1156.385961] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: f132230d-0fba-4293-b940-85cc7ef1cad1] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1156.875651] env[62753]: DEBUG nova.network.neutron [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance_info_cache with network_info: [{"id": "ad95cf04-2066-4452-b160-65cf3cddbe87", "address": "fa:16:3e:eb:5e:01", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad95cf04-20", "ovs_interfaceid": "ad95cf04-2066-4452-b160-65cf3cddbe87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1156.888669] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 4323eb5a-b12c-4e31-be90-fcaf6d3ad5d3] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1157.379118] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1157.380215] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e3aef98-3fb1-45c0-841a-38258b1bf810 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.388599] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c3b684d-9d1b-4933-a398-34bd246e1e1f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1157.391537] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 046fc3dc-a187-4fda-bc66-345e1226f83d] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1157.894815] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 3bc05fdb-81be-4764-b9d9-01acd125b020] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1158.398460] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 41476adb-95e5-4617-b178-9c981d13c43d] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1158.471549] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1158.471795] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1cb40893-48d4-4b9f-9f2a-ee94947bc195 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1158.479789] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1158.479789] env[62753]: value = "task-1332888" [ 1158.479789] env[62753]: _type = "Task" [ 1158.479789] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1158.487571] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332888, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1158.902065] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._cleanup_incomplete_migrations {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1158.902065] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Cleaning up deleted instances with incomplete migration {{(pid=62753) _cleanup_incomplete_migrations /opt/stack/nova/nova/compute/manager.py:11266}} [ 1158.988319] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332888, 'name': PowerOffVM_Task, 'duration_secs': 0.315503} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1158.988579] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1158.989264] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1158.989488] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1158.989649] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1158.989835] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1158.990026] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1158.990193] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1158.990416] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1158.990593] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1158.990771] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1158.990942] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1158.991135] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1158.995997] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-969fe42f-2538-4d74-b1ee-47d6e25d7ff3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.011860] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1159.011860] env[62753]: value = "task-1332889" [ 1159.011860] env[62753]: _type = "Task" [ 1159.011860] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.019117] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332889, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1159.522086] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332889, 'name': ReconfigVM_Task, 'duration_secs': 0.142207} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1159.522894] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-95d1aa46-c64c-4216-b386-0f106c5986e0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.543267] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1159.543532] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1159.543693] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1159.543882] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1159.544475] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1159.544475] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1159.544475] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1159.544627] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1159.544810] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1159.544982] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1159.545199] env[62753]: DEBUG nova.virt.hardware [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1159.546016] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-4095b26b-16ef-4882-a547-f2ecd79fe9fb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1159.551682] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1159.551682] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523332f5-9f6b-a54a-dfbc-d04c0e1bd2d9" [ 1159.551682] env[62753]: _type = "Task" [ 1159.551682] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1159.559156] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523332f5-9f6b-a54a-dfbc-d04c0e1bd2d9, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.061766] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]523332f5-9f6b-a54a-dfbc-d04c0e1bd2d9, 'name': SearchDatastore_Task, 'duration_secs': 0.006676} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.066990] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Reconfiguring VM instance instance-00000067 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1160.067284] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-6947dd57-427a-4535-8a96-b48cdae93784 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.086550] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1160.086550] env[62753]: value = "task-1332890" [ 1160.086550] env[62753]: _type = "Task" [ 1160.086550] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.098997] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332890, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.404080] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1160.404255] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Starting heal instance info cache {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1160.404327] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Rebuilding the list of instances to heal {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1160.596437] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332890, 'name': ReconfigVM_Task, 'duration_secs': 0.250907} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1160.596722] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Reconfigured VM instance instance-00000067 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1160.597484] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-47c23363-87f3-41db-88c4-d9ef5b79a227 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.621197] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Reconfiguring VM instance instance-00000067 to attach disk [datastore1] ca01cd3f-7f0e-4c61-a704-74f9945792b4/ca01cd3f-7f0e-4c61-a704-74f9945792b4.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1160.621514] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-fbbe4a1b-9e73-4661-aa48-afa1966555c9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1160.638586] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1160.638586] env[62753]: value = "task-1332891" [ 1160.638586] env[62753]: _type = "Task" [ 1160.638586] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1160.646290] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332891, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1160.909326] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1160.909493] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquired lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1160.909551] env[62753]: DEBUG nova.network.neutron [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Forcefully refreshing network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1160.909687] env[62753]: DEBUG nova.objects.instance [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lazy-loading 'info_cache' on Instance uuid ca01cd3f-7f0e-4c61-a704-74f9945792b4 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1161.150556] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332891, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1161.649536] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332891, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.149665] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332891, 'name': ReconfigVM_Task, 'duration_secs': 1.294419} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.149982] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Reconfigured VM instance instance-00000067 to attach disk [datastore1] ca01cd3f-7f0e-4c61-a704-74f9945792b4/ca01cd3f-7f0e-4c61-a704-74f9945792b4.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1162.150845] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b168683b-94c0-496b-b65d-88374dfe77a1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.170524] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a0130dc-d5de-4c81-8fc7-2e331d94d467 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.190145] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a897a6e9-b075-4027-a032-d7565dfc2e55 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.209921] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cf6c2b6-cd60-46a1-aaf8-fdafdd576add {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.216448] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1162.216672] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-95c78ece-3968-4385-8f23-b5d78afaf0c0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1162.222847] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1162.222847] env[62753]: value = "task-1332892" [ 1162.222847] env[62753]: _type = "Task" [ 1162.222847] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1162.230977] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332892, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1162.624952] env[62753]: DEBUG nova.network.neutron [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance_info_cache with network_info: [{"id": "ad95cf04-2066-4452-b160-65cf3cddbe87", "address": "fa:16:3e:eb:5e:01", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapad95cf04-20", "ovs_interfaceid": "ad95cf04-2066-4452-b160-65cf3cddbe87", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1162.738454] env[62753]: DEBUG oslo_vmware.api [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332892, 'name': PowerOnVM_Task, 'duration_secs': 0.360208} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1162.738669] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1163.127379] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Releasing lock "refresh_cache-ca01cd3f-7f0e-4c61-a704-74f9945792b4" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1163.127604] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updated the network info_cache for instance {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1163.127802] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1163.127973] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._sync_power_states {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1163.632757] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Getting list of instances from cluster (obj){ [ 1163.632757] env[62753]: value = "domain-c8" [ 1163.632757] env[62753]: _type = "ClusterComputeResource" [ 1163.632757] env[62753]: } {{(pid=62753) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2116}} [ 1163.633914] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c70bca7-04d2-45d5-afb5-d2064d49dfdf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.650249] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Got total of 7 instances {{(pid=62753) list_instances /opt/stack/nova/nova/virt/vmwareapi/vmops.py:2125}} [ 1163.650429] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Triggering sync for uuid ca01cd3f-7f0e-4c61-a704-74f9945792b4 {{(pid=62753) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1163.650601] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Triggering sync for uuid 22cea165-cfde-403f-a0e2-82a86f5baa51 {{(pid=62753) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1163.650754] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Triggering sync for uuid a2e7fa98-566e-40f1-8da7-9318de89c14c {{(pid=62753) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1163.650916] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Triggering sync for uuid 51790f51-c464-48cc-9891-41cae5a8f63e {{(pid=62753) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1163.651146] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Triggering sync for uuid ec0b07b8-1cb9-4911-8b03-efb466a21888 {{(pid=62753) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1163.651313] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Triggering sync for uuid 21e6cf03-e528-4289-8813-86020a4e8d2a {{(pid=62753) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1163.651463] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Triggering sync for uuid 6a97171d-7ea2-4581-8707-b6be33f1e8ed {{(pid=62753) _sync_power_states /opt/stack/nova/nova/compute/manager.py:10338}} [ 1163.651851] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.652132] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.652325] env[62753]: INFO nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] During sync_power_state the instance has a pending task (resize_reverting). Skip. [ 1163.652504] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1163.652704] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "22cea165-cfde-403f-a0e2-82a86f5baa51" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.652891] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.653163] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "a2e7fa98-566e-40f1-8da7-9318de89c14c" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.653351] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "a2e7fa98-566e-40f1-8da7-9318de89c14c" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.653599] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "51790f51-c464-48cc-9891-41cae5a8f63e" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.653811] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "51790f51-c464-48cc-9891-41cae5a8f63e" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.654050] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.654236] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.654462] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "21e6cf03-e528-4289-8813-86020a4e8d2a" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.654643] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "21e6cf03-e528-4289-8813-86020a4e8d2a" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.654863] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "6a97171d-7ea2-4581-8707-b6be33f1e8ed" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1163.655096] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "6a97171d-7ea2-4581-8707-b6be33f1e8ed" acquired by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1163.655930] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-338719c2-7694-4fb7-88cc-1d535a28ceca {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.659114] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d3b279d5-516e-419e-9b12-d3b3049507d0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.661840] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0ca1d132-444f-4e11-a189-edc453034cb4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.664524] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3de6492-3e83-4785-818d-0ce2ad4a0767 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.667315] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a392abd-0a49-46c2-9903-857f904241de {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.670812] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6da99f7-7c5f-47c6-be6a-eb6b25a537ff {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1163.778038] env[62753]: INFO nova.compute.manager [None req-5345b247-e74f-4363-b2e5-44908b42c6ad tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance to original state: 'active' [ 1164.188065] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "a2e7fa98-566e-40f1-8da7-9318de89c14c" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.534s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.190693] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.538s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.191019] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.537s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.193489] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "51790f51-c464-48cc-9891-41cae5a8f63e" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.540s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.193793] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "21e6cf03-e528-4289-8813-86020a4e8d2a" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.539s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.196200] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "6a97171d-7ea2-4581-8707-b6be33f1e8ed" "released" by "nova.compute.manager.ComputeManager._sync_power_states.._sync..query_driver_power_state_and_sync" :: held 0.541s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.857485] env[62753]: DEBUG oslo_concurrency.lockutils [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.857924] env[62753]: DEBUG oslo_concurrency.lockutils [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.857999] env[62753]: DEBUG oslo_concurrency.lockutils [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1164.858176] env[62753]: DEBUG oslo_concurrency.lockutils [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1164.858325] env[62753]: DEBUG oslo_concurrency.lockutils [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1164.861023] env[62753]: INFO nova.compute.manager [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Terminating instance [ 1164.862388] env[62753]: DEBUG nova.compute.manager [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1164.862611] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1164.862853] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-deb72e2e-33c1-4c91-8b89-d7af2603f903 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1164.869486] env[62753]: DEBUG oslo_vmware.api [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1164.869486] env[62753]: value = "task-1332893" [ 1164.869486] env[62753]: _type = "Task" [ 1164.869486] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1164.878356] env[62753]: DEBUG oslo_vmware.api [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332893, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.379379] env[62753]: DEBUG oslo_vmware.api [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332893, 'name': PowerOffVM_Task, 'duration_secs': 0.178704} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.379643] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1165.379844] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Volume detach. Driver type: vmdk {{(pid=62753) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1165.380046] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284694', 'volume_id': 'ea8a3d93-7c93-4b76-9194-d4539fc86180', 'name': 'volume-ea8a3d93-7c93-4b76-9194-d4539fc86180', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'ca01cd3f-7f0e-4c61-a704-74f9945792b4', 'attached_at': '2024-10-21T20:36:22.000000', 'detached_at': '', 'volume_id': 'ea8a3d93-7c93-4b76-9194-d4539fc86180', 'serial': 'ea8a3d93-7c93-4b76-9194-d4539fc86180'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1165.380798] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4687f621-1349-4088-9d43-f5972b8159a9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.400809] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82f558db-476b-48e6-acef-3db4a93e8363 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.406715] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5e80fb1c-ac68-4972-9876-bbce4e029cf1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.425714] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d94e4d15-b0d1-4705-a352-17e6a5f75a9a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.438918] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] The volume has not been displaced from its original location: [datastore2] volume-ea8a3d93-7c93-4b76-9194-d4539fc86180/volume-ea8a3d93-7c93-4b76-9194-d4539fc86180.vmdk. No consolidation needed. {{(pid=62753) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1165.444046] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Reconfiguring VM instance instance-00000067 to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1165.444312] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-640c32e4-15f2-4d0e-8c09-e1889f953265 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.461469] env[62753]: DEBUG oslo_vmware.api [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1165.461469] env[62753]: value = "task-1332894" [ 1165.461469] env[62753]: _type = "Task" [ 1165.461469] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.468778] env[62753]: DEBUG oslo_vmware.api [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332894, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1165.970937] env[62753]: DEBUG oslo_vmware.api [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332894, 'name': ReconfigVM_Task, 'duration_secs': 0.197104} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1165.971349] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Reconfigured VM instance instance-00000067 to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1165.976334] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-003976ae-9b71-42e0-8b12-8007b9d0be98 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1165.990579] env[62753]: DEBUG oslo_vmware.api [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1165.990579] env[62753]: value = "task-1332895" [ 1165.990579] env[62753]: _type = "Task" [ 1165.990579] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1165.998815] env[62753]: DEBUG oslo_vmware.api [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332895, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.344349] env[62753]: DEBUG oslo_concurrency.lockutils [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1166.344593] env[62753]: DEBUG oslo_concurrency.lockutils [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1166.500761] env[62753]: DEBUG oslo_vmware.api [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332895, 'name': ReconfigVM_Task, 'duration_secs': 0.121048} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1166.501098] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284694', 'volume_id': 'ea8a3d93-7c93-4b76-9194-d4539fc86180', 'name': 'volume-ea8a3d93-7c93-4b76-9194-d4539fc86180', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'attaching', 'instance': 'ca01cd3f-7f0e-4c61-a704-74f9945792b4', 'attached_at': '2024-10-21T20:36:22.000000', 'detached_at': '', 'volume_id': 'ea8a3d93-7c93-4b76-9194-d4539fc86180', 'serial': 'ea8a3d93-7c93-4b76-9194-d4539fc86180'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1166.501392] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1166.502145] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-28eb9579-8ae7-4195-a80d-4c02fe96227f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.508237] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1166.508452] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-25573627-d0ab-4b25-9f07-d4662ede7a02 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.568012] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1166.568285] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1166.568483] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Deleting the datastore file [datastore1] ca01cd3f-7f0e-4c61-a704-74f9945792b4 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1166.568747] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-22f05053-5f98-4d98-a95a-ec2c643b2ab0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1166.574694] env[62753]: DEBUG oslo_vmware.api [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1166.574694] env[62753]: value = "task-1332897" [ 1166.574694] env[62753]: _type = "Task" [ 1166.574694] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1166.582922] env[62753]: DEBUG oslo_vmware.api [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332897, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1166.847245] env[62753]: DEBUG nova.compute.utils [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1167.085050] env[62753]: DEBUG oslo_vmware.api [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332897, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.139901} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1167.085397] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1167.085561] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1167.085784] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1167.085987] env[62753]: INFO nova.compute.manager [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Took 2.22 seconds to destroy the instance on the hypervisor. [ 1167.086256] env[62753]: DEBUG oslo.service.loopingcall [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1167.086448] env[62753]: DEBUG nova.compute.manager [-] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1167.086544] env[62753]: DEBUG nova.network.neutron [-] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1167.324620] env[62753]: DEBUG oslo_concurrency.lockutils [None req-37d64795-f772-4a51-8106-7f0266cb4418 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "22cea165-cfde-403f-a0e2-82a86f5baa51" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1167.325249] env[62753]: DEBUG oslo_concurrency.lockutils [None req-37d64795-f772-4a51-8106-7f0266cb4418 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51" acquired by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1167.325249] env[62753]: INFO nova.compute.manager [None req-37d64795-f772-4a51-8106-7f0266cb4418 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Rebooting instance [ 1167.350173] env[62753]: DEBUG oslo_concurrency.lockutils [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.005s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1167.564585] env[62753]: DEBUG nova.compute.manager [req-f356b9b7-5c69-4968-b446-4888fae2e350 req-0ddcfee9-0d61-4b7a-bd52-93f301d185e1 service nova] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Received event network-vif-deleted-ad95cf04-2066-4452-b160-65cf3cddbe87 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1167.564797] env[62753]: INFO nova.compute.manager [req-f356b9b7-5c69-4968-b446-4888fae2e350 req-0ddcfee9-0d61-4b7a-bd52-93f301d185e1 service nova] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Neutron deleted interface ad95cf04-2066-4452-b160-65cf3cddbe87; detaching it from the instance and deleting it from the info cache [ 1167.564964] env[62753]: DEBUG nova.network.neutron [req-f356b9b7-5c69-4968-b446-4888fae2e350 req-0ddcfee9-0d61-4b7a-bd52-93f301d185e1 service nova] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1167.847436] env[62753]: DEBUG oslo_concurrency.lockutils [None req-37d64795-f772-4a51-8106-7f0266cb4418 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1167.847622] env[62753]: DEBUG oslo_concurrency.lockutils [None req-37d64795-f772-4a51-8106-7f0266cb4418 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1167.847799] env[62753]: DEBUG nova.network.neutron [None req-37d64795-f772-4a51-8106-7f0266cb4418 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1168.044622] env[62753]: DEBUG nova.network.neutron [-] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1168.067667] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-def9c1c1-b64b-4e96-8c46-404bfe69c5b0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.079788] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bb99daef-4838-4aa5-ad57-208aa504d7be {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.106190] env[62753]: DEBUG nova.compute.manager [req-f356b9b7-5c69-4968-b446-4888fae2e350 req-0ddcfee9-0d61-4b7a-bd52-93f301d185e1 service nova] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Detach interface failed, port_id=ad95cf04-2066-4452-b160-65cf3cddbe87, reason: Instance ca01cd3f-7f0e-4c61-a704-74f9945792b4 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1168.427515] env[62753]: DEBUG oslo_concurrency.lockutils [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1168.427742] env[62753]: DEBUG oslo_concurrency.lockutils [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1168.428015] env[62753]: INFO nova.compute.manager [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Attaching volume dc9f6473-9f05-4ca4-a314-4a1175d096bb to /dev/sdb [ 1168.457652] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-837dcdaa-08eb-48cf-b5dd-6fe252c5bd92 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.466545] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa07ce8a-1914-401e-a0e6-36fcf95e724d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1168.479083] env[62753]: DEBUG nova.virt.block_device [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Updating existing volume attachment record: 6c387175-fb04-4f84-b60b-f6039158299a {{(pid=62753) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1168.550598] env[62753]: INFO nova.compute.manager [-] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Took 1.46 seconds to deallocate network for instance. [ 1168.573014] env[62753]: DEBUG nova.network.neutron [None req-37d64795-f772-4a51-8106-7f0266cb4418 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance_info_cache with network_info: [{"id": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "address": "fa:16:3e:88:6a:1d", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25e4c4c8-9d", "ovs_interfaceid": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1169.077384] env[62753]: DEBUG oslo_concurrency.lockutils [None req-37d64795-f772-4a51-8106-7f0266cb4418 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1169.079598] env[62753]: DEBUG nova.compute.manager [None req-37d64795-f772-4a51-8106-7f0266cb4418 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1169.080479] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e6ecba0d-ade7-4114-9610-52c483b86f20 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1169.090269] env[62753]: INFO nova.compute.manager [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Took 0.54 seconds to detach 1 volumes for instance. [ 1169.599412] env[62753]: DEBUG oslo_concurrency.lockutils [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1169.599412] env[62753]: DEBUG oslo_concurrency.lockutils [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1169.599412] env[62753]: DEBUG oslo_concurrency.lockutils [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1169.625278] env[62753]: INFO nova.scheduler.client.report [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Deleted allocations for instance ca01cd3f-7f0e-4c61-a704-74f9945792b4 [ 1170.098997] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54e92d4-241a-4f97-b59e-ccf82cf3dadb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.106926] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-37d64795-f772-4a51-8106-7f0266cb4418 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Doing hard reboot of VM {{(pid=62753) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1063}} [ 1170.107189] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ResetVM_Task with opID=oslo.vmware-9f626dd8-dbc5-43cf-b99c-b506f57b5733 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.112700] env[62753]: DEBUG oslo_vmware.api [None req-37d64795-f772-4a51-8106-7f0266cb4418 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1170.112700] env[62753]: value = "task-1332901" [ 1170.112700] env[62753]: _type = "Task" [ 1170.112700] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1170.121219] env[62753]: DEBUG oslo_vmware.api [None req-37d64795-f772-4a51-8106-7f0266cb4418 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332901, 'name': ResetVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1170.133963] env[62753]: DEBUG oslo_concurrency.lockutils [None req-afa8e280-ecc1-4611-b006-7b67dd28014c tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "ca01cd3f-7f0e-4c61-a704-74f9945792b4" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.275s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1170.624198] env[62753]: DEBUG oslo_vmware.api [None req-37d64795-f772-4a51-8106-7f0266cb4418 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332901, 'name': ResetVM_Task, 'duration_secs': 0.09149} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1170.624552] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-37d64795-f772-4a51-8106-7f0266cb4418 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Did hard reboot of VM {{(pid=62753) reboot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1067}} [ 1170.624602] env[62753]: DEBUG nova.compute.manager [None req-37d64795-f772-4a51-8106-7f0266cb4418 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1170.625350] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a9401c18-1415-4a39-a489-2c8730d73543 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1170.878186] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "2e784985-7045-448e-a650-f6692bb69d94" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1170.878426] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "2e784985-7045-448e-a650-f6692bb69d94" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.137083] env[62753]: DEBUG oslo_concurrency.lockutils [None req-37d64795-f772-4a51-8106-7f0266cb4418 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51" "released" by "nova.compute.manager.ComputeManager.reboot_instance..do_reboot_instance" :: held 3.812s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1171.382054] env[62753]: DEBUG nova.compute.manager [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1171.902973] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1171.903266] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1171.904800] env[62753]: INFO nova.compute.claims [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1172.994463] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bad0ba54-9e4c-4eeb-88d1-70dab3d252d4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.001627] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-764e17f3-880a-4729-8542-528481d489dd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.032056] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Volume attach. Driver type: vmdk {{(pid=62753) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1173.032293] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284700', 'volume_id': 'dc9f6473-9f05-4ca4-a314-4a1175d096bb', 'name': 'volume-dc9f6473-9f05-4ca4-a314-4a1175d096bb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ec0b07b8-1cb9-4911-8b03-efb466a21888', 'attached_at': '', 'detached_at': '', 'volume_id': 'dc9f6473-9f05-4ca4-a314-4a1175d096bb', 'serial': 'dc9f6473-9f05-4ca4-a314-4a1175d096bb'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1173.033018] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d325fa3d-a6fb-49d9-82ef-80a56e5cbe57 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.035763] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-abc87491-37cc-4a95-9d75-d0ec252a7a3e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.052335] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63c0570c-2972-479b-bf3c-702517ca5f6f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.056475] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-98a2f48d-e17f-4537-b8da-6f4de63dfc15 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.068123] env[62753]: DEBUG nova.compute.provider_tree [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1173.089198] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Reconfiguring VM instance instance-0000006b to attach disk [datastore1] volume-dc9f6473-9f05-4ca4-a314-4a1175d096bb/volume-dc9f6473-9f05-4ca4-a314-4a1175d096bb.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1173.090118] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aa470841-eb56-4205-92df-157678e44037 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.107362] env[62753]: DEBUG oslo_vmware.api [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1173.107362] env[62753]: value = "task-1332903" [ 1173.107362] env[62753]: _type = "Task" [ 1173.107362] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.115050] env[62753]: DEBUG oslo_vmware.api [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332903, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1173.570891] env[62753]: DEBUG nova.scheduler.client.report [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1173.617206] env[62753]: DEBUG oslo_vmware.api [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332903, 'name': ReconfigVM_Task, 'duration_secs': 0.362438} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1173.617478] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Reconfigured VM instance instance-0000006b to attach disk [datastore1] volume-dc9f6473-9f05-4ca4-a314-4a1175d096bb/volume-dc9f6473-9f05-4ca4-a314-4a1175d096bb.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1173.622256] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d1f54c29-6a2b-46ed-bec2-7a564c098bf8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1173.636635] env[62753]: DEBUG oslo_vmware.api [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1173.636635] env[62753]: value = "task-1332904" [ 1173.636635] env[62753]: _type = "Task" [ 1173.636635] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1173.643868] env[62753]: DEBUG oslo_vmware.api [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332904, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1174.076841] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.173s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1174.077443] env[62753]: DEBUG nova.compute.manager [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1174.146926] env[62753]: DEBUG oslo_vmware.api [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332904, 'name': ReconfigVM_Task, 'duration_secs': 0.137743} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1174.147184] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284700', 'volume_id': 'dc9f6473-9f05-4ca4-a314-4a1175d096bb', 'name': 'volume-dc9f6473-9f05-4ca4-a314-4a1175d096bb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ec0b07b8-1cb9-4911-8b03-efb466a21888', 'attached_at': '', 'detached_at': '', 'volume_id': 'dc9f6473-9f05-4ca4-a314-4a1175d096bb', 'serial': 'dc9f6473-9f05-4ca4-a314-4a1175d096bb'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1174.582912] env[62753]: DEBUG nova.compute.utils [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1174.584433] env[62753]: DEBUG nova.compute.manager [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1174.584612] env[62753]: DEBUG nova.network.neutron [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1174.654816] env[62753]: DEBUG nova.policy [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'af45e909b0c84662a8cfdf92896e5fe5', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '196457eaa51643c38b4ef4fc6fbbcba3', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1174.916064] env[62753]: DEBUG nova.network.neutron [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Successfully created port: b81e545c-6410-4acd-8777-1c9a2ef2b7eb {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1175.088749] env[62753]: DEBUG nova.compute.manager [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1175.187993] env[62753]: DEBUG nova.objects.instance [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lazy-loading 'flavor' on Instance uuid ec0b07b8-1cb9-4911-8b03-efb466a21888 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1175.694425] env[62753]: DEBUG oslo_concurrency.lockutils [None req-10982b2d-c31e-4758-a652-ceb4b3f1cd80 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.266s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1175.930140] env[62753]: INFO nova.compute.manager [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Rescuing [ 1175.930494] env[62753]: DEBUG oslo_concurrency.lockutils [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1175.930677] env[62753]: DEBUG oslo_concurrency.lockutils [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1175.930863] env[62753]: DEBUG nova.network.neutron [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1176.004813] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "21e6cf03-e528-4289-8813-86020a4e8d2a" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.005070] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "21e6cf03-e528-4289-8813-86020a4e8d2a" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.098021] env[62753]: DEBUG nova.compute.manager [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1176.122666] env[62753]: DEBUG nova.virt.hardware [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1176.122958] env[62753]: DEBUG nova.virt.hardware [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1176.123179] env[62753]: DEBUG nova.virt.hardware [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1176.123409] env[62753]: DEBUG nova.virt.hardware [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1176.123573] env[62753]: DEBUG nova.virt.hardware [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1176.123728] env[62753]: DEBUG nova.virt.hardware [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1176.123947] env[62753]: DEBUG nova.virt.hardware [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1176.124123] env[62753]: DEBUG nova.virt.hardware [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1176.124298] env[62753]: DEBUG nova.virt.hardware [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1176.124466] env[62753]: DEBUG nova.virt.hardware [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1176.124643] env[62753]: DEBUG nova.virt.hardware [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1176.125507] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c5729ffc-7f76-41f3-9780-8e534b348805 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.133538] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4a26af6f-aaf0-4e33-878b-f7afea998187 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1176.312153] env[62753]: DEBUG nova.compute.manager [req-e6ebcb0f-5cd8-4fe9-9a43-a25a0e4b4e03 req-b81b8b1c-6931-4d85-be4f-4b0c776b577b service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Received event network-vif-plugged-b81e545c-6410-4acd-8777-1c9a2ef2b7eb {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1176.312455] env[62753]: DEBUG oslo_concurrency.lockutils [req-e6ebcb0f-5cd8-4fe9-9a43-a25a0e4b4e03 req-b81b8b1c-6931-4d85-be4f-4b0c776b577b service nova] Acquiring lock "2e784985-7045-448e-a650-f6692bb69d94-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1176.312593] env[62753]: DEBUG oslo_concurrency.lockutils [req-e6ebcb0f-5cd8-4fe9-9a43-a25a0e4b4e03 req-b81b8b1c-6931-4d85-be4f-4b0c776b577b service nova] Lock "2e784985-7045-448e-a650-f6692bb69d94-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1176.312763] env[62753]: DEBUG oslo_concurrency.lockutils [req-e6ebcb0f-5cd8-4fe9-9a43-a25a0e4b4e03 req-b81b8b1c-6931-4d85-be4f-4b0c776b577b service nova] Lock "2e784985-7045-448e-a650-f6692bb69d94-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1176.312934] env[62753]: DEBUG nova.compute.manager [req-e6ebcb0f-5cd8-4fe9-9a43-a25a0e4b4e03 req-b81b8b1c-6931-4d85-be4f-4b0c776b577b service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] No waiting events found dispatching network-vif-plugged-b81e545c-6410-4acd-8777-1c9a2ef2b7eb {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1176.313112] env[62753]: WARNING nova.compute.manager [req-e6ebcb0f-5cd8-4fe9-9a43-a25a0e4b4e03 req-b81b8b1c-6931-4d85-be4f-4b0c776b577b service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Received unexpected event network-vif-plugged-b81e545c-6410-4acd-8777-1c9a2ef2b7eb for instance with vm_state building and task_state spawning. [ 1176.508882] env[62753]: DEBUG nova.compute.utils [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1176.663252] env[62753]: DEBUG nova.network.neutron [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Updating instance_info_cache with network_info: [{"id": "8600276a-3346-44f1-a760-df6caca2e51e", "address": "fa:16:3e:bf:ad:12", "network": {"id": "939bb8aa-f8d4-44cb-960f-c5a5fbcf99c8", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1149502111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dd81bdb01ff4c39a4959f4af2acf61a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8600276a-33", "ovs_interfaceid": "8600276a-3346-44f1-a760-df6caca2e51e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1176.873707] env[62753]: DEBUG nova.network.neutron [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Successfully updated port: b81e545c-6410-4acd-8777-1c9a2ef2b7eb {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1176.897529] env[62753]: DEBUG nova.compute.manager [req-c654c6a7-0bff-4ee8-8f28-c70115894826 req-2953971c-a2a2-400c-a1b7-96e3aa57b42b service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Received event network-changed-b81e545c-6410-4acd-8777-1c9a2ef2b7eb {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1176.897756] env[62753]: DEBUG nova.compute.manager [req-c654c6a7-0bff-4ee8-8f28-c70115894826 req-2953971c-a2a2-400c-a1b7-96e3aa57b42b service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Refreshing instance network info cache due to event network-changed-b81e545c-6410-4acd-8777-1c9a2ef2b7eb. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1176.897962] env[62753]: DEBUG oslo_concurrency.lockutils [req-c654c6a7-0bff-4ee8-8f28-c70115894826 req-2953971c-a2a2-400c-a1b7-96e3aa57b42b service nova] Acquiring lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1176.898151] env[62753]: DEBUG oslo_concurrency.lockutils [req-c654c6a7-0bff-4ee8-8f28-c70115894826 req-2953971c-a2a2-400c-a1b7-96e3aa57b42b service nova] Acquired lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1176.898291] env[62753]: DEBUG nova.network.neutron [req-c654c6a7-0bff-4ee8-8f28-c70115894826 req-2953971c-a2a2-400c-a1b7-96e3aa57b42b service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Refreshing network info cache for port b81e545c-6410-4acd-8777-1c9a2ef2b7eb {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1177.011768] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "21e6cf03-e528-4289-8813-86020a4e8d2a" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.006s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1177.165901] env[62753]: DEBUG oslo_concurrency.lockutils [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Releasing lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1177.380688] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1177.429274] env[62753]: DEBUG nova.network.neutron [req-c654c6a7-0bff-4ee8-8f28-c70115894826 req-2953971c-a2a2-400c-a1b7-96e3aa57b42b service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1177.503815] env[62753]: DEBUG nova.network.neutron [req-c654c6a7-0bff-4ee8-8f28-c70115894826 req-2953971c-a2a2-400c-a1b7-96e3aa57b42b service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1177.692462] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1177.692683] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a23001f4-6f57-4331-84ba-e839b5630ff1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1177.701141] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1177.701141] env[62753]: value = "task-1332905" [ 1177.701141] env[62753]: _type = "Task" [ 1177.701141] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1177.714362] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332905, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.007082] env[62753]: DEBUG oslo_concurrency.lockutils [req-c654c6a7-0bff-4ee8-8f28-c70115894826 req-2953971c-a2a2-400c-a1b7-96e3aa57b42b service nova] Releasing lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1178.007396] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.007562] env[62753]: DEBUG nova.network.neutron [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1178.085319] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "21e6cf03-e528-4289-8813-86020a4e8d2a" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1178.085566] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "21e6cf03-e528-4289-8813-86020a4e8d2a" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1178.085797] env[62753]: INFO nova.compute.manager [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Attaching volume 31c0173a-45d2-44ed-8315-1dfae366307d to /dev/sdb [ 1178.119953] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-696fbc82-27b3-4b0b-ac40-2f331a6c6a7d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.127065] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ae3009bf-7e42-4bdf-97d5-4bfd01454715 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.140299] env[62753]: DEBUG nova.virt.block_device [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Updating existing volume attachment record: 3e306cf6-948b-414f-8abb-67049e71071b {{(pid=62753) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1178.211813] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332905, 'name': PowerOffVM_Task, 'duration_secs': 0.198796} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.212109] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1178.212858] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f18afeb4-eaea-41c4-98e8-7072a2452362 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.232735] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31d4107c-d5af-4c02-a236-3b55098fe9ae {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.274166] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1178.274650] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-1e6b1cea-3264-4617-96d0-91b293e5b6c4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.284172] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1178.284172] env[62753]: value = "task-1332906" [ 1178.284172] env[62753]: _type = "Task" [ 1178.284172] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.291226] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] VM already powered off {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1178.291436] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1178.291688] env[62753]: DEBUG oslo_concurrency.lockutils [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1178.292275] env[62753]: DEBUG oslo_concurrency.lockutils [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1178.292275] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1178.292402] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-77407808-085b-4fad-9a9f-ebbb980d2272 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.301543] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1178.301726] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1178.302429] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-da53e0f4-5f1e-44a5-8abc-1e5468519746 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.307341] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1178.307341] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52410042-40d6-1d05-acb8-d5a6028eefab" [ 1178.307341] env[62753]: _type = "Task" [ 1178.307341] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.314561] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52410042-40d6-1d05-acb8-d5a6028eefab, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1178.567847] env[62753]: DEBUG nova.network.neutron [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1178.695834] env[62753]: DEBUG nova.network.neutron [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Updating instance_info_cache with network_info: [{"id": "b81e545c-6410-4acd-8777-1c9a2ef2b7eb", "address": "fa:16:3e:12:cc:32", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb81e545c-64", "ovs_interfaceid": "b81e545c-6410-4acd-8777-1c9a2ef2b7eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1178.818060] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52410042-40d6-1d05-acb8-d5a6028eefab, 'name': SearchDatastore_Task, 'duration_secs': 0.008818} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1178.818246] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-cb03edfa-a545-4702-be0e-7c4ed71637bf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1178.823019] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1178.823019] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525a0fe8-87fa-48c9-f095-4b031d447e27" [ 1178.823019] env[62753]: _type = "Task" [ 1178.823019] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1178.830398] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525a0fe8-87fa-48c9-f095-4b031d447e27, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.198671] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.199018] env[62753]: DEBUG nova.compute.manager [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Instance network_info: |[{"id": "b81e545c-6410-4acd-8777-1c9a2ef2b7eb", "address": "fa:16:3e:12:cc:32", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb81e545c-64", "ovs_interfaceid": "b81e545c-6410-4acd-8777-1c9a2ef2b7eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1179.199538] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:12:cc:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b81e545c-6410-4acd-8777-1c9a2ef2b7eb', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1179.206969] env[62753]: DEBUG oslo.service.loopingcall [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1179.207198] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1179.207475] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-02e3ab01-953b-47e2-8ca4-f6fcc2ab2d23 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.227548] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1179.227548] env[62753]: value = "task-1332908" [ 1179.227548] env[62753]: _type = "Task" [ 1179.227548] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.235793] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332908, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.333738] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525a0fe8-87fa-48c9-f095-4b031d447e27, 'name': SearchDatastore_Task, 'duration_secs': 0.008493} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.333941] env[62753]: DEBUG oslo_concurrency.lockutils [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1179.334217] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Copying virtual disk from [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] ec0b07b8-1cb9-4911-8b03-efb466a21888/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk. {{(pid=62753) disk_copy /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:281}} [ 1179.334505] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-f749caa9-3521-4874-ad85-2129a9c38fab {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.340808] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1179.340808] env[62753]: value = "task-1332909" [ 1179.340808] env[62753]: _type = "Task" [ 1179.340808] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.348240] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332909, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.738944] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332908, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1179.851886] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332909, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.444663} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1179.852160] env[62753]: INFO nova.virt.vmwareapi.ds_util [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Copied virtual disk from [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] ec0b07b8-1cb9-4911-8b03-efb466a21888/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk. [ 1179.852899] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c6b10a42-9bf6-4fca-8946-dd5a10b71345 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.879050] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Reconfiguring VM instance instance-0000006b to attach disk [datastore2] ec0b07b8-1cb9-4911-8b03-efb466a21888/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1179.879151] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-9b97a08a-87f1-4b2e-a491-b98d377d3de9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1179.896715] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1179.896715] env[62753]: value = "task-1332910" [ 1179.896715] env[62753]: _type = "Task" [ 1179.896715] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1179.907219] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332910, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.238057] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332908, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.406098] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332910, 'name': ReconfigVM_Task, 'duration_secs': 0.327821} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.406415] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Reconfigured VM instance instance-0000006b to attach disk [datastore2] ec0b07b8-1cb9-4911-8b03-efb466a21888/a7541f15-bad6-4ea8-95ce-3499a4f01dda-rescue.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1180.407275] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c72f5581-95ae-4c60-b7f6-89aa4f9c0fbc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.433995] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-456f49f8-ea6f-451d-a49a-dba57759e75b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.452210] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1180.452210] env[62753]: value = "task-1332912" [ 1180.452210] env[62753]: _type = "Task" [ 1180.452210] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.460898] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332912, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.738696] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332908, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.949689] env[62753]: INFO nova.compute.manager [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Rebuilding instance [ 1180.962641] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332912, 'name': ReconfigVM_Task, 'duration_secs': 0.17589} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1180.962980] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1180.963290] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d7de601f-c5ae-47f3-bb09-28fd510e8248 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1180.969674] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1180.969674] env[62753]: value = "task-1332913" [ 1180.969674] env[62753]: _type = "Task" [ 1180.969674] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1180.982740] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332913, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1180.994758] env[62753]: DEBUG nova.compute.manager [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1180.995566] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-05cdcc27-000a-4959-90c8-6ec95d674fd6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.240201] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332908, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.480734] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332913, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.506414] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1181.506668] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9aafd7b0-45df-484f-9e36-e2db3099af7e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1181.513700] env[62753]: DEBUG oslo_vmware.api [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Waiting for the task: (returnval){ [ 1181.513700] env[62753]: value = "task-1332914" [ 1181.513700] env[62753]: _type = "Task" [ 1181.513700] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1181.521349] env[62753]: DEBUG oslo_vmware.api [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332914, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.740741] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332908, 'name': CreateVM_Task} progress is 25%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1181.980930] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332913, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.022682] env[62753]: DEBUG oslo_vmware.api [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332914, 'name': PowerOffVM_Task, 'duration_secs': 0.239204} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.022954] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1182.023659] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1182.023907] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-bf8e71ed-2dbf-459c-8178-8d6de2e5e300 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.029593] env[62753]: DEBUG oslo_vmware.api [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Waiting for the task: (returnval){ [ 1182.029593] env[62753]: value = "task-1332915" [ 1182.029593] env[62753]: _type = "Task" [ 1182.029593] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.036575] env[62753]: DEBUG oslo_vmware.api [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332915, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.242070] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332908, 'name': CreateVM_Task, 'duration_secs': 2.602047} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.242070] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1182.242556] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1182.242723] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.243019] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1182.243280] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-3c65d7a8-73df-4760-ab6d-1256f310b073 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.248144] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1182.248144] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]527091d1-dcac-5555-5b37-4e74c3266e53" [ 1182.248144] env[62753]: _type = "Task" [ 1182.248144] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.256181] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]527091d1-dcac-5555-5b37-4e74c3266e53, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.481610] env[62753]: DEBUG oslo_vmware.api [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332913, 'name': PowerOnVM_Task, 'duration_secs': 1.013967} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.481874] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1182.484590] env[62753]: DEBUG nova.compute.manager [None req-eb065a5d-4918-416e-b62a-c139b853a7ef tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1182.485337] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4594d24c-8de5-4351-9360-6a1a08df4d2a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.539738] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] VM already powered off {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1182.539893] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Volume detach. Driver type: vmdk {{(pid=62753) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1182.540109] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284692', 'volume_id': 'e055f16d-248a-4dca-97d1-ce25b21061b5', 'name': 'volume-e055f16d-248a-4dca-97d1-ce25b21061b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6a97171d-7ea2-4581-8707-b6be33f1e8ed', 'attached_at': '', 'detached_at': '', 'volume_id': 'e055f16d-248a-4dca-97d1-ce25b21061b5', 'serial': 'e055f16d-248a-4dca-97d1-ce25b21061b5'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1182.540862] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9782eb45-8b6d-411b-9cb6-a3934b353d3f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.558374] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da81133a-491a-4beb-b145-5582a75472c1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.565634] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-90c0a74f-48c4-4f15-bd95-f49fb727040c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.582312] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-24fb1918-9205-40fc-a513-ef91918b0182 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.596454] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] The volume has not been displaced from its original location: [datastore1] volume-e055f16d-248a-4dca-97d1-ce25b21061b5/volume-e055f16d-248a-4dca-97d1-ce25b21061b5.vmdk. No consolidation needed. {{(pid=62753) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1182.601536] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Reconfiguring VM instance instance-0000006d to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1182.601781] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-aee3f152-a56c-47b8-8b13-4a0385c34a93 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.618350] env[62753]: DEBUG oslo_vmware.api [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Waiting for the task: (returnval){ [ 1182.618350] env[62753]: value = "task-1332916" [ 1182.618350] env[62753]: _type = "Task" [ 1182.618350] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.625465] env[62753]: DEBUG oslo_vmware.api [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332916, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.684413] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Volume attach. Driver type: vmdk {{(pid=62753) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1182.684652] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284701', 'volume_id': '31c0173a-45d2-44ed-8315-1dfae366307d', 'name': 'volume-31c0173a-45d2-44ed-8315-1dfae366307d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '21e6cf03-e528-4289-8813-86020a4e8d2a', 'attached_at': '', 'detached_at': '', 'volume_id': '31c0173a-45d2-44ed-8315-1dfae366307d', 'serial': '31c0173a-45d2-44ed-8315-1dfae366307d'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1182.685535] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ea988942-8ab2-46a9-a78b-7b0f2c35659f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.701437] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-38e49c0a-6759-494d-a94c-48629861c649 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.725406] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Reconfiguring VM instance instance-0000006c to attach disk [datastore1] volume-31c0173a-45d2-44ed-8315-1dfae366307d/volume-31c0173a-45d2-44ed-8315-1dfae366307d.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1182.725723] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-ccd2b9a3-0fbd-42f6-bede-d6f13d390330 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.743883] env[62753]: DEBUG oslo_vmware.api [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1182.743883] env[62753]: value = "task-1332917" [ 1182.743883] env[62753]: _type = "Task" [ 1182.743883] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.754897] env[62753]: DEBUG oslo_vmware.api [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332917, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1182.760767] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]527091d1-dcac-5555-5b37-4e74c3266e53, 'name': SearchDatastore_Task, 'duration_secs': 0.010986} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1182.761087] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1182.761343] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1182.761579] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1182.761729] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1182.761910] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1182.762192] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-151b4f1e-73c3-48fe-b696-2d6512d60859 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.770425] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1182.770625] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1182.771391] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-5de3bbc2-9ab3-4e7a-9687-2fe571e9e289 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1182.776635] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1182.776635] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5270e9fd-bd45-472f-b80a-1bd2b3a74e9a" [ 1182.776635] env[62753]: _type = "Task" [ 1182.776635] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1182.786095] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5270e9fd-bd45-472f-b80a-1bd2b3a74e9a, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.128453] env[62753]: DEBUG oslo_vmware.api [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332916, 'name': ReconfigVM_Task, 'duration_secs': 0.202957} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.128690] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Reconfigured VM instance instance-0000006d to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1183.133185] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c9a64096-ff0b-454e-a97a-077d5156872d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.148519] env[62753]: DEBUG oslo_vmware.api [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Waiting for the task: (returnval){ [ 1183.148519] env[62753]: value = "task-1332918" [ 1183.148519] env[62753]: _type = "Task" [ 1183.148519] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.158192] env[62753]: DEBUG oslo_vmware.api [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332918, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.253946] env[62753]: DEBUG oslo_vmware.api [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332917, 'name': ReconfigVM_Task, 'duration_secs': 0.362637} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.254298] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Reconfigured VM instance instance-0000006c to attach disk [datastore1] volume-31c0173a-45d2-44ed-8315-1dfae366307d/volume-31c0173a-45d2-44ed-8315-1dfae366307d.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1183.259121] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-e8a798b0-e100-43b2-ace8-09893f3175be {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.273133] env[62753]: DEBUG oslo_vmware.api [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1183.273133] env[62753]: value = "task-1332919" [ 1183.273133] env[62753]: _type = "Task" [ 1183.273133] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.282882] env[62753]: DEBUG oslo_vmware.api [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332919, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.288227] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5270e9fd-bd45-472f-b80a-1bd2b3a74e9a, 'name': SearchDatastore_Task, 'duration_secs': 0.009859} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.288807] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-e5d7c7d2-aea8-4911-b32c-c21c1c76e8dd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.293676] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1183.293676] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a945e0-13b8-5172-9f22-457c4d9fe6e8" [ 1183.293676] env[62753]: _type = "Task" [ 1183.293676] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.301397] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a945e0-13b8-5172-9f22-457c4d9fe6e8, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.658827] env[62753]: DEBUG oslo_vmware.api [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332918, 'name': ReconfigVM_Task, 'duration_secs': 0.104087} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.659093] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284692', 'volume_id': 'e055f16d-248a-4dca-97d1-ce25b21061b5', 'name': 'volume-e055f16d-248a-4dca-97d1-ce25b21061b5', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '6a97171d-7ea2-4581-8707-b6be33f1e8ed', 'attached_at': '', 'detached_at': '', 'volume_id': 'e055f16d-248a-4dca-97d1-ce25b21061b5', 'serial': 'e055f16d-248a-4dca-97d1-ce25b21061b5'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1183.659389] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1183.660158] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be1d1340-bafe-4abb-a913-28d7892ce348 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.666147] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1183.666372] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-0b8c9486-12a7-4506-a284-86e6f4492732 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.727720] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1183.727966] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1183.728195] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Deleting the datastore file [datastore1] 6a97171d-7ea2-4581-8707-b6be33f1e8ed {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1183.728473] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-2ee473b9-0120-4ff1-8aec-ed1d79680211 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.734855] env[62753]: DEBUG oslo_vmware.api [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Waiting for the task: (returnval){ [ 1183.734855] env[62753]: value = "task-1332921" [ 1183.734855] env[62753]: _type = "Task" [ 1183.734855] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.742480] env[62753]: DEBUG oslo_vmware.api [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332921, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.782885] env[62753]: DEBUG oslo_vmware.api [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332919, 'name': ReconfigVM_Task, 'duration_secs': 0.154404} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.783297] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284701', 'volume_id': '31c0173a-45d2-44ed-8315-1dfae366307d', 'name': 'volume-31c0173a-45d2-44ed-8315-1dfae366307d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '21e6cf03-e528-4289-8813-86020a4e8d2a', 'attached_at': '', 'detached_at': '', 'volume_id': '31c0173a-45d2-44ed-8315-1dfae366307d', 'serial': '31c0173a-45d2-44ed-8315-1dfae366307d'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1183.803681] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a945e0-13b8-5172-9f22-457c4d9fe6e8, 'name': SearchDatastore_Task, 'duration_secs': 0.009906} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1183.803938] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1183.804219] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 2e784985-7045-448e-a650-f6692bb69d94/2e784985-7045-448e-a650-f6692bb69d94.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1183.804481] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-a0a1af71-6a42-4e02-bcfa-8e73173c8516 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1183.810414] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1183.810414] env[62753]: value = "task-1332922" [ 1183.810414] env[62753]: _type = "Task" [ 1183.810414] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1183.817884] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332922, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1183.946178] env[62753]: INFO nova.compute.manager [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Unrescuing [ 1183.946499] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1183.946656] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquired lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1183.946833] env[62753]: DEBUG nova.network.neutron [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1184.247031] env[62753]: DEBUG oslo_vmware.api [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Task: {'id': task-1332921, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.074717} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.247375] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1184.247604] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1184.247808] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1184.299810] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Volume detach. Driver type: vmdk {{(pid=62753) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1184.300196] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-309cc39d-4949-4529-b2b9-ec220a28f88f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.308899] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b8777cf-8af0-4ad1-ac29-bb2c43022f27 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.330276] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332922, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.438921} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.330539] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 2e784985-7045-448e-a650-f6692bb69d94/2e784985-7045-448e-a650-f6692bb69d94.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1184.330772] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1184.331584] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-39e878b0-1c06-45db-b196-dabc7f1dbf11 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.342979] env[62753]: ERROR nova.compute.manager [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Failed to detach volume e055f16d-248a-4dca-97d1-ce25b21061b5 from /dev/sda: nova.exception.InstanceNotFound: Instance 6a97171d-7ea2-4581-8707-b6be33f1e8ed could not be found. [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Traceback (most recent call last): [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] self.driver.rebuild(**kwargs) [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] raise NotImplementedError() [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] NotImplementedError [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] During handling of the above exception, another exception occurred: [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Traceback (most recent call last): [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] self.driver.detach_volume(context, old_connection_info, [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 559, in detach_volume [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] return self._volumeops.detach_volume(connection_info, instance) [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] self._detach_volume_vmdk(connection_info, instance) [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] stable_ref.fetch_moref(session) [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] nova.exception.InstanceNotFound: Instance 6a97171d-7ea2-4581-8707-b6be33f1e8ed could not be found. [ 1184.342979] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] [ 1184.346704] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1184.346704] env[62753]: value = "task-1332923" [ 1184.346704] env[62753]: _type = "Task" [ 1184.346704] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.355673] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332923, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1184.475267] env[62753]: DEBUG nova.compute.utils [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Build of instance 6a97171d-7ea2-4581-8707-b6be33f1e8ed aborted: Failed to rebuild volume backed instance. {{(pid=62753) notify_about_instance_usage /opt/stack/nova/nova/compute/utils.py:430}} [ 1184.477617] env[62753]: ERROR nova.compute.manager [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Setting instance vm_state to ERROR: nova.exception.BuildAbortException: Build of instance 6a97171d-7ea2-4581-8707-b6be33f1e8ed aborted: Failed to rebuild volume backed instance. [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Traceback (most recent call last): [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/compute/manager.py", line 4142, in _do_rebuild_instance [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] self.driver.rebuild(**kwargs) [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/virt/driver.py", line 493, in rebuild [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] raise NotImplementedError() [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] NotImplementedError [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] During handling of the above exception, another exception occurred: [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Traceback (most recent call last): [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/compute/manager.py", line 3600, in _rebuild_volume_backed_instance [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] self._detach_root_volume(context, instance, root_bdm) [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/compute/manager.py", line 3579, in _detach_root_volume [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] with excutils.save_and_reraise_exception(): [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 227, in __exit__ [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] self.force_reraise() [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/data/venv/lib/python3.10/site-packages/oslo_utils/excutils.py", line 200, in force_reraise [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] raise self.value [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/compute/manager.py", line 3565, in _detach_root_volume [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] self.driver.detach_volume(context, old_connection_info, [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/virt/vmwareapi/driver.py", line 559, in detach_volume [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] return self._volumeops.detach_volume(connection_info, instance) [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 649, in detach_volume [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] self._detach_volume_vmdk(connection_info, instance) [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/virt/vmwareapi/volumeops.py", line 569, in _detach_volume_vmdk [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] vm_ref = vm_util.get_vm_ref(self._session, instance) [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1145, in get_vm_ref [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] stable_ref.fetch_moref(session) [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/virt/vmwareapi/vm_util.py", line 1136, in fetch_moref [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] raise exception.InstanceNotFound(instance_id=self._uuid) [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] nova.exception.InstanceNotFound: Instance 6a97171d-7ea2-4581-8707-b6be33f1e8ed could not be found. [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] During handling of the above exception, another exception occurred: [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Traceback (most recent call last): [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/compute/manager.py", line 10865, in _error_out_instance_on_exception [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] yield [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/compute/manager.py", line 3868, in rebuild_instance [ 1184.477617] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] self._do_rebuild_instance_with_claim( [ 1184.478799] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/compute/manager.py", line 3954, in _do_rebuild_instance_with_claim [ 1184.478799] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] self._do_rebuild_instance( [ 1184.478799] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/compute/manager.py", line 4146, in _do_rebuild_instance [ 1184.478799] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] self._rebuild_default_impl(**kwargs) [ 1184.478799] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/compute/manager.py", line 3723, in _rebuild_default_impl [ 1184.478799] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] self._rebuild_volume_backed_instance( [ 1184.478799] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] File "/opt/stack/nova/nova/compute/manager.py", line 3615, in _rebuild_volume_backed_instance [ 1184.478799] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] raise exception.BuildAbortException( [ 1184.478799] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] nova.exception.BuildAbortException: Build of instance 6a97171d-7ea2-4581-8707-b6be33f1e8ed aborted: Failed to rebuild volume backed instance. [ 1184.478799] env[62753]: ERROR nova.compute.manager [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] [ 1184.826756] env[62753]: DEBUG nova.objects.instance [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lazy-loading 'flavor' on Instance uuid 21e6cf03-e528-4289-8813-86020a4e8d2a {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1184.845931] env[62753]: DEBUG nova.network.neutron [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Updating instance_info_cache with network_info: [{"id": "8600276a-3346-44f1-a760-df6caca2e51e", "address": "fa:16:3e:bf:ad:12", "network": {"id": "939bb8aa-f8d4-44cb-960f-c5a5fbcf99c8", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1149502111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dd81bdb01ff4c39a4959f4af2acf61a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8600276a-33", "ovs_interfaceid": "8600276a-3346-44f1-a760-df6caca2e51e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1184.856725] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332923, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060995} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1184.856987] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1184.857780] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc40b7ee-8927-41ee-9a53-7072bf098bb0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.879393] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Reconfiguring VM instance instance-0000006e to attach disk [datastore1] 2e784985-7045-448e-a650-f6692bb69d94/2e784985-7045-448e-a650-f6692bb69d94.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1184.879853] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-bb59cf87-a811-4686-949a-1173d849756b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1184.899180] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1184.899180] env[62753]: value = "task-1332924" [ 1184.899180] env[62753]: _type = "Task" [ 1184.899180] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1184.906733] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332924, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.331828] env[62753]: DEBUG oslo_concurrency.lockutils [None req-1a6c2b99-2c8e-4a94-9c56-5fdb6221ba11 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "21e6cf03-e528-4289-8813-86020a4e8d2a" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.246s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1185.348910] env[62753]: DEBUG oslo_concurrency.lockutils [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Releasing lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1185.349519] env[62753]: DEBUG nova.objects.instance [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lazy-loading 'flavor' on Instance uuid ec0b07b8-1cb9-4911-8b03-efb466a21888 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1185.408254] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332924, 'name': ReconfigVM_Task, 'duration_secs': 0.288997} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.408517] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Reconfigured VM instance instance-0000006e to attach disk [datastore1] 2e784985-7045-448e-a650-f6692bb69d94/2e784985-7045-448e-a650-f6692bb69d94.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1185.409764] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4e7b8f4e-ce96-469c-ac14-6a6a923da02d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.415014] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1185.415014] env[62753]: value = "task-1332925" [ 1185.415014] env[62753]: _type = "Task" [ 1185.415014] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.422244] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332925, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.565675] env[62753]: DEBUG oslo_concurrency.lockutils [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "21e6cf03-e528-4289-8813-86020a4e8d2a" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1185.565944] env[62753]: DEBUG oslo_concurrency.lockutils [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "21e6cf03-e528-4289-8813-86020a4e8d2a" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1185.854785] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f7ca73de-6143-41cf-bfce-84df8c6d60c5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.879571] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1185.879905] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-6e14f855-1d23-4cf7-bbea-c312ce406950 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.887069] env[62753]: DEBUG oslo_vmware.api [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1185.887069] env[62753]: value = "task-1332926" [ 1185.887069] env[62753]: _type = "Task" [ 1185.887069] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.895918] env[62753]: DEBUG oslo_vmware.api [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332926, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1185.924993] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332925, 'name': Rename_Task, 'duration_secs': 0.138949} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1185.925279] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1185.925524] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4aa9a87a-8706-4d73-98aa-7003a6dedc82 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1185.931748] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1185.931748] env[62753]: value = "task-1332927" [ 1185.931748] env[62753]: _type = "Task" [ 1185.931748] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1185.939876] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332927, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.069563] env[62753]: INFO nova.compute.manager [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Detaching volume 31c0173a-45d2-44ed-8315-1dfae366307d [ 1186.103814] env[62753]: INFO nova.virt.block_device [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Attempting to driver detach volume 31c0173a-45d2-44ed-8315-1dfae366307d from mountpoint /dev/sdb [ 1186.104239] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Volume detach. Driver type: vmdk {{(pid=62753) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1186.104434] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284701', 'volume_id': '31c0173a-45d2-44ed-8315-1dfae366307d', 'name': 'volume-31c0173a-45d2-44ed-8315-1dfae366307d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '21e6cf03-e528-4289-8813-86020a4e8d2a', 'attached_at': '', 'detached_at': '', 'volume_id': '31c0173a-45d2-44ed-8315-1dfae366307d', 'serial': '31c0173a-45d2-44ed-8315-1dfae366307d'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1186.105408] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc5679b5-5d31-4e7c-b82b-555719a064b6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.130669] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ae18691-d47f-4fc6-ac6d-223bf969133c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.137765] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77941a46-3d8a-48c9-8314-f1abe71a1db6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.158263] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa148302-1a48-4975-8aef-e10fcdcbf016 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.172794] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] The volume has not been displaced from its original location: [datastore1] volume-31c0173a-45d2-44ed-8315-1dfae366307d/volume-31c0173a-45d2-44ed-8315-1dfae366307d.vmdk. No consolidation needed. {{(pid=62753) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1186.178139] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Reconfiguring VM instance instance-0000006c to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1186.178448] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-472c49a4-0bf6-40ea-bb92-3face124bfff {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.196958] env[62753]: DEBUG oslo_vmware.api [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1186.196958] env[62753]: value = "task-1332928" [ 1186.196958] env[62753]: _type = "Task" [ 1186.196958] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.204672] env[62753]: DEBUG oslo_vmware.api [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332928, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.400963] env[62753]: DEBUG oslo_vmware.api [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332926, 'name': PowerOffVM_Task, 'duration_secs': 0.253733} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.401296] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1186.406556] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Reconfiguring VM instance instance-0000006b to detach disk 2002 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1186.406843] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-77758acc-6fcd-4f2b-82f4-327889230950 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.425673] env[62753]: DEBUG oslo_vmware.api [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1186.425673] env[62753]: value = "task-1332929" [ 1186.425673] env[62753]: _type = "Task" [ 1186.425673] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.433496] env[62753]: DEBUG oslo_vmware.api [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332929, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.440530] env[62753]: DEBUG oslo_vmware.api [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332927, 'name': PowerOnVM_Task, 'duration_secs': 0.472778} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.440793] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1186.440996] env[62753]: INFO nova.compute.manager [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Took 10.34 seconds to spawn the instance on the hypervisor. [ 1186.441196] env[62753]: DEBUG nova.compute.manager [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1186.441914] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7e68eca1-4301-4787-bf87-95aba560cb01 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.494680] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.494953] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1186.578336] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-43ac34d8-b447-4677-b1f6-627e01fb5f98 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.585997] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-420e57a7-8a76-4224-8d80-e3e3e8522040 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.615914] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4b344ee5-c1e5-4ac2-8611-f7e3c54863b2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.623255] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-372b78da-8f64-40da-917b-a3fcda78eade {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.636484] env[62753]: DEBUG nova.compute.provider_tree [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1186.706847] env[62753]: DEBUG oslo_vmware.api [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332928, 'name': ReconfigVM_Task, 'duration_secs': 0.215215} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.707231] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Reconfigured VM instance instance-0000006c to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1186.711972] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c5e91ac4-9f43-4331-a6ff-7fbfcb4bdc4b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.726081] env[62753]: DEBUG oslo_vmware.api [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1186.726081] env[62753]: value = "task-1332930" [ 1186.726081] env[62753]: _type = "Task" [ 1186.726081] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.733623] env[62753]: DEBUG oslo_vmware.api [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332930, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1186.882179] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Acquiring lock "6a97171d-7ea2-4581-8707-b6be33f1e8ed" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.882475] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Lock "6a97171d-7ea2-4581-8707-b6be33f1e8ed" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1186.882720] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Acquiring lock "6a97171d-7ea2-4581-8707-b6be33f1e8ed-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1186.882921] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Lock "6a97171d-7ea2-4581-8707-b6be33f1e8ed-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1186.883232] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Lock "6a97171d-7ea2-4581-8707-b6be33f1e8ed-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1186.885695] env[62753]: INFO nova.compute.manager [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Terminating instance [ 1186.888050] env[62753]: DEBUG nova.compute.manager [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1186.888166] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ca65c8fd-a667-4b4f-99ea-24c7d9e0f060 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.897583] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ca1ee7d1-810f-4bca-a5db-c073f3ad362d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.927894] env[62753]: WARNING nova.virt.vmwareapi.driver [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Instance does not exists. Proceeding to delete instance properties on datastore: nova.exception.InstanceNotFound: Instance 6a97171d-7ea2-4581-8707-b6be33f1e8ed could not be found. [ 1186.928102] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1186.931051] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9a9cbc5f-e6bd-4388-ac3e-599bd151d3c2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.938388] env[62753]: DEBUG oslo_vmware.api [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332929, 'name': ReconfigVM_Task, 'duration_secs': 0.245175} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1186.939381] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Reconfigured VM instance instance-0000006b to detach disk 2002 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1186.939575] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1186.939836] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c360d9be-e67f-4894-a6c5-bae2c2527069 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.943671] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12c9282c-d38b-45c3-8dbf-84f9765acbd9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1186.959531] env[62753]: INFO nova.compute.manager [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Took 15.07 seconds to build instance. [ 1186.962045] env[62753]: DEBUG oslo_vmware.api [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1186.962045] env[62753]: value = "task-1332931" [ 1186.962045] env[62753]: _type = "Task" [ 1186.962045] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1186.973922] env[62753]: WARNING nova.virt.vmwareapi.vmops [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Instance does not exist on backend: nova.exception.InstanceNotFound: Instance 6a97171d-7ea2-4581-8707-b6be33f1e8ed could not be found. [ 1186.974164] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1186.974360] env[62753]: INFO nova.compute.manager [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Took 0.09 seconds to destroy the instance on the hypervisor. [ 1186.974602] env[62753]: DEBUG oslo.service.loopingcall [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1186.978279] env[62753]: DEBUG nova.compute.manager [-] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1186.978391] env[62753]: DEBUG nova.network.neutron [-] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1186.980499] env[62753]: DEBUG oslo_vmware.api [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332931, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.139956] env[62753]: DEBUG nova.scheduler.client.report [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1187.236922] env[62753]: DEBUG oslo_vmware.api [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332930, 'name': ReconfigVM_Task, 'duration_secs': 0.135417} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1187.237311] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284701', 'volume_id': '31c0173a-45d2-44ed-8315-1dfae366307d', 'name': 'volume-31c0173a-45d2-44ed-8315-1dfae366307d', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '21e6cf03-e528-4289-8813-86020a4e8d2a', 'attached_at': '', 'detached_at': '', 'volume_id': '31c0173a-45d2-44ed-8315-1dfae366307d', 'serial': '31c0173a-45d2-44ed-8315-1dfae366307d'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1187.354271] env[62753]: DEBUG nova.compute.manager [req-1857cb8e-82d8-4c77-aca6-99ab8a1c2799 req-5ded8d73-c25c-492e-90b6-6e4d79ef8932 service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Received event network-changed-b81e545c-6410-4acd-8777-1c9a2ef2b7eb {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1187.354528] env[62753]: DEBUG nova.compute.manager [req-1857cb8e-82d8-4c77-aca6-99ab8a1c2799 req-5ded8d73-c25c-492e-90b6-6e4d79ef8932 service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Refreshing instance network info cache due to event network-changed-b81e545c-6410-4acd-8777-1c9a2ef2b7eb. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1187.354803] env[62753]: DEBUG oslo_concurrency.lockutils [req-1857cb8e-82d8-4c77-aca6-99ab8a1c2799 req-5ded8d73-c25c-492e-90b6-6e4d79ef8932 service nova] Acquiring lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1187.355014] env[62753]: DEBUG oslo_concurrency.lockutils [req-1857cb8e-82d8-4c77-aca6-99ab8a1c2799 req-5ded8d73-c25c-492e-90b6-6e4d79ef8932 service nova] Acquired lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1187.355196] env[62753]: DEBUG nova.network.neutron [req-1857cb8e-82d8-4c77-aca6-99ab8a1c2799 req-5ded8d73-c25c-492e-90b6-6e4d79ef8932 service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Refreshing network info cache for port b81e545c-6410-4acd-8777-1c9a2ef2b7eb {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1187.461907] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5a3495ef-6889-4e36-8c91-63c8511ac666 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "2e784985-7045-448e-a650-f6692bb69d94" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 16.583s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.472652] env[62753]: DEBUG oslo_vmware.api [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332931, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1187.537356] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1187.645571] env[62753]: DEBUG oslo_concurrency.lockutils [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.150s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1187.645571] env[62753]: INFO nova.compute.manager [None req-0b6ddcc1-41e0-4eba-b2f9-2436d8115c1a tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Successfully reverted task state from rebuilding on failure for instance. [ 1187.797196] env[62753]: DEBUG nova.objects.instance [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lazy-loading 'flavor' on Instance uuid 21e6cf03-e528-4289-8813-86020a4e8d2a {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1187.976195] env[62753]: DEBUG oslo_vmware.api [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332931, 'name': PowerOnVM_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1188.010327] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.010564] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1188.010830] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62753) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1188.056129] env[62753]: DEBUG nova.network.neutron [-] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.099705] env[62753]: DEBUG nova.network.neutron [req-1857cb8e-82d8-4c77-aca6-99ab8a1c2799 req-5ded8d73-c25c-492e-90b6-6e4d79ef8932 service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Updated VIF entry in instance network info cache for port b81e545c-6410-4acd-8777-1c9a2ef2b7eb. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1188.100097] env[62753]: DEBUG nova.network.neutron [req-1857cb8e-82d8-4c77-aca6-99ab8a1c2799 req-5ded8d73-c25c-492e-90b6-6e4d79ef8932 service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Updating instance_info_cache with network_info: [{"id": "b81e545c-6410-4acd-8777-1c9a2ef2b7eb", "address": "fa:16:3e:12:cc:32", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb81e545c-64", "ovs_interfaceid": "b81e545c-6410-4acd-8777-1c9a2ef2b7eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1188.473181] env[62753]: DEBUG oslo_vmware.api [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332931, 'name': PowerOnVM_Task, 'duration_secs': 1.025865} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1188.473485] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1188.473638] env[62753]: DEBUG nova.compute.manager [None req-6f1b88aa-0558-442c-87c5-5d4161a006b8 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1188.474408] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a412f4d1-62f8-4baf-93aa-050e14f4db9b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1188.562198] env[62753]: INFO nova.compute.manager [-] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Took 1.58 seconds to deallocate network for instance. [ 1188.602801] env[62753]: DEBUG oslo_concurrency.lockutils [req-1857cb8e-82d8-4c77-aca6-99ab8a1c2799 req-5ded8d73-c25c-492e-90b6-6e4d79ef8932 service nova] Releasing lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1188.803705] env[62753]: DEBUG oslo_concurrency.lockutils [None req-65e90df8-edfb-4f7a-ad05-a8179370b4a6 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "21e6cf03-e528-4289-8813-86020a4e8d2a" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.238s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.009837] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1189.113803] env[62753]: INFO nova.compute.manager [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Took 0.55 seconds to detach 1 volumes for instance. [ 1189.118230] env[62753]: DEBUG nova.compute.manager [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Deleting volume: e055f16d-248a-4dca-97d1-ce25b21061b5 {{(pid=62753) _cleanup_volumes /opt/stack/nova/nova/compute/manager.py:3247}} [ 1189.382997] env[62753]: DEBUG nova.compute.manager [req-d7d2b99f-d4f7-4766-86a9-6f46c3bb3baf req-cc750220-2d13-4091-9e8b-0b0ab2452d59 service nova] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Received event network-vif-deleted-0dd198e5-3992-4c1f-939a-16be88877273 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1189.670897] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1189.671279] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.671395] env[62753]: DEBUG nova.objects.instance [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Lazy-loading 'resources' on Instance uuid 6a97171d-7ea2-4581-8707-b6be33f1e8ed {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1189.834740] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "21e6cf03-e528-4289-8813-86020a4e8d2a" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1189.835214] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "21e6cf03-e528-4289-8813-86020a4e8d2a" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.835597] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "21e6cf03-e528-4289-8813-86020a4e8d2a-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1189.835845] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "21e6cf03-e528-4289-8813-86020a4e8d2a-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1189.836078] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "21e6cf03-e528-4289-8813-86020a4e8d2a-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1189.838924] env[62753]: INFO nova.compute.manager [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Terminating instance [ 1189.841385] env[62753]: DEBUG nova.compute.manager [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1189.841672] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1189.842858] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e34db563-7c2c-4050-acaf-5fd2792c113e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.852941] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1189.853259] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-a17cbf44-4e2a-4906-86e9-bac79165ce7a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1189.859312] env[62753]: DEBUG oslo_vmware.api [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1189.859312] env[62753]: value = "task-1332933" [ 1189.859312] env[62753]: _type = "Task" [ 1189.859312] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1189.867300] env[62753]: DEBUG oslo_vmware.api [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332933, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.010635] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1190.260344] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-64c1806f-454d-4bad-aff4-5e4a3a4a9da5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.268866] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3eb458db-d6dd-4b23-ae3c-e56197a18387 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.299509] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-459f258f-46c9-434c-9822-ae13aed8b75b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.307214] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a54bee39-245f-42c2-90e5-faa8ecd6fd98 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.320490] env[62753]: DEBUG nova.compute.provider_tree [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1190.369661] env[62753]: DEBUG oslo_vmware.api [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332933, 'name': PowerOffVM_Task, 'duration_secs': 0.219434} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1190.369945] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1190.370135] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1190.370395] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-3ceb4896-c758-43fa-9f20-53cc3849cdc3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.443922] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1190.443922] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1190.444115] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Deleting the datastore file [datastore2] 21e6cf03-e528-4289-8813-86020a4e8d2a {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1190.444426] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-f236a7c0-0104-45ea-b9d5-1e8280a7c4cc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1190.453117] env[62753]: DEBUG oslo_vmware.api [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1190.453117] env[62753]: value = "task-1332935" [ 1190.453117] env[62753]: _type = "Task" [ 1190.453117] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1190.461243] env[62753]: DEBUG oslo_vmware.api [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332935, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1190.824171] env[62753]: DEBUG nova.scheduler.client.report [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1190.963776] env[62753]: DEBUG oslo_vmware.api [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332935, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1191.005822] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1191.009465] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1191.329700] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.658s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.416215] env[62753]: DEBUG nova.compute.manager [req-ca06d230-ca6c-4caf-8d86-cdc76211e3c1 req-b200b589-7de3-4718-a28c-947c3151fcc3 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Received event network-changed-8600276a-3346-44f1-a760-df6caca2e51e {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1191.416424] env[62753]: DEBUG nova.compute.manager [req-ca06d230-ca6c-4caf-8d86-cdc76211e3c1 req-b200b589-7de3-4718-a28c-947c3151fcc3 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Refreshing instance network info cache due to event network-changed-8600276a-3346-44f1-a760-df6caca2e51e. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1191.416643] env[62753]: DEBUG oslo_concurrency.lockutils [req-ca06d230-ca6c-4caf-8d86-cdc76211e3c1 req-b200b589-7de3-4718-a28c-947c3151fcc3 service nova] Acquiring lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1191.416790] env[62753]: DEBUG oslo_concurrency.lockutils [req-ca06d230-ca6c-4caf-8d86-cdc76211e3c1 req-b200b589-7de3-4718-a28c-947c3151fcc3 service nova] Acquired lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1191.416956] env[62753]: DEBUG nova.network.neutron [req-ca06d230-ca6c-4caf-8d86-cdc76211e3c1 req-b200b589-7de3-4718-a28c-947c3151fcc3 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Refreshing network info cache for port 8600276a-3346-44f1-a760-df6caca2e51e {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1191.465287] env[62753]: DEBUG oslo_vmware.api [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332935, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.761637} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1191.465614] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1191.465864] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1191.466067] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1191.466276] env[62753]: INFO nova.compute.manager [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Took 1.62 seconds to destroy the instance on the hypervisor. [ 1191.466526] env[62753]: DEBUG oslo.service.loopingcall [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1191.466714] env[62753]: DEBUG nova.compute.manager [-] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1191.466809] env[62753]: DEBUG nova.network.neutron [-] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1191.512159] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.512381] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.512548] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1191.512696] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62753) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1191.513570] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d46060c5-3f0a-4cf4-ae4e-bfb941f57db4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.521609] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d5ec61be-7cf0-4913-916b-37e809d50f21 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.536764] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29623e2e-90d8-4dfa-8907-7f633af4cf15 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.543247] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-305b1934-bcd1-475a-a071-5c17e4a2e144 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1191.573959] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180320MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=62753) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1191.573959] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1191.574253] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1191.848365] env[62753]: DEBUG oslo_concurrency.lockutils [None req-3485134d-aa60-48ae-8f4f-8e3addbad806 tempest-ServerActionsV293TestJSON-1092600123 tempest-ServerActionsV293TestJSON-1092600123-project-member] Lock "6a97171d-7ea2-4581-8707-b6be33f1e8ed" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 4.966s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1192.139798] env[62753]: DEBUG nova.network.neutron [req-ca06d230-ca6c-4caf-8d86-cdc76211e3c1 req-b200b589-7de3-4718-a28c-947c3151fcc3 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Updated VIF entry in instance network info cache for port 8600276a-3346-44f1-a760-df6caca2e51e. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1192.140288] env[62753]: DEBUG nova.network.neutron [req-ca06d230-ca6c-4caf-8d86-cdc76211e3c1 req-b200b589-7de3-4718-a28c-947c3151fcc3 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Updating instance_info_cache with network_info: [{"id": "8600276a-3346-44f1-a760-df6caca2e51e", "address": "fa:16:3e:bf:ad:12", "network": {"id": "939bb8aa-f8d4-44cb-960f-c5a5fbcf99c8", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1149502111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dd81bdb01ff4c39a4959f4af2acf61a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8600276a-33", "ovs_interfaceid": "8600276a-3346-44f1-a760-df6caca2e51e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1192.383536] env[62753]: DEBUG nova.network.neutron [-] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1192.601373] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 22cea165-cfde-403f-a0e2-82a86f5baa51 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1192.601474] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance a2e7fa98-566e-40f1-8da7-9318de89c14c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1192.601606] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 51790f51-c464-48cc-9891-41cae5a8f63e actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1192.601728] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance ec0b07b8-1cb9-4911-8b03-efb466a21888 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1192.601845] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 21e6cf03-e528-4289-8813-86020a4e8d2a actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1192.601961] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 2e784985-7045-448e-a650-f6692bb69d94 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1192.602162] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Total usable vcpus: 48, total allocated vcpus: 6 {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1192.602300] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1664MB phys_disk=200GB used_disk=6GB total_vcpus=48 used_vcpus=6 pci_stats=[] {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1192.642851] env[62753]: DEBUG oslo_concurrency.lockutils [req-ca06d230-ca6c-4caf-8d86-cdc76211e3c1 req-b200b589-7de3-4718-a28c-947c3151fcc3 service nova] Releasing lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1192.643134] env[62753]: DEBUG nova.compute.manager [req-ca06d230-ca6c-4caf-8d86-cdc76211e3c1 req-b200b589-7de3-4718-a28c-947c3151fcc3 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Received event network-changed-8600276a-3346-44f1-a760-df6caca2e51e {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1192.643314] env[62753]: DEBUG nova.compute.manager [req-ca06d230-ca6c-4caf-8d86-cdc76211e3c1 req-b200b589-7de3-4718-a28c-947c3151fcc3 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Refreshing instance network info cache due to event network-changed-8600276a-3346-44f1-a760-df6caca2e51e. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1192.643530] env[62753]: DEBUG oslo_concurrency.lockutils [req-ca06d230-ca6c-4caf-8d86-cdc76211e3c1 req-b200b589-7de3-4718-a28c-947c3151fcc3 service nova] Acquiring lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1192.643677] env[62753]: DEBUG oslo_concurrency.lockutils [req-ca06d230-ca6c-4caf-8d86-cdc76211e3c1 req-b200b589-7de3-4718-a28c-947c3151fcc3 service nova] Acquired lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1192.643853] env[62753]: DEBUG nova.network.neutron [req-ca06d230-ca6c-4caf-8d86-cdc76211e3c1 req-b200b589-7de3-4718-a28c-947c3151fcc3 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Refreshing network info cache for port 8600276a-3346-44f1-a760-df6caca2e51e {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1192.681186] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3d947da-10d9-4305-85a3-3c50ae4c489b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.689251] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7638cd24-b810-4ba8-905d-d350eeab69ef {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.718824] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8a003526-296c-4b8e-a004-89834357d8c0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.727026] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2c06da0b-82dd-4e56-8853-a631e582bcdf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1192.740806] env[62753]: DEBUG nova.compute.provider_tree [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1192.885934] env[62753]: INFO nova.compute.manager [-] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Took 1.42 seconds to deallocate network for instance. [ 1193.244388] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1193.345695] env[62753]: DEBUG nova.network.neutron [req-ca06d230-ca6c-4caf-8d86-cdc76211e3c1 req-b200b589-7de3-4718-a28c-947c3151fcc3 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Updated VIF entry in instance network info cache for port 8600276a-3346-44f1-a760-df6caca2e51e. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1193.346088] env[62753]: DEBUG nova.network.neutron [req-ca06d230-ca6c-4caf-8d86-cdc76211e3c1 req-b200b589-7de3-4718-a28c-947c3151fcc3 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Updating instance_info_cache with network_info: [{"id": "8600276a-3346-44f1-a760-df6caca2e51e", "address": "fa:16:3e:bf:ad:12", "network": {"id": "939bb8aa-f8d4-44cb-960f-c5a5fbcf99c8", "bridge": "br-int", "label": "tempest-ServerRescueNegativeTestJSON-1149502111-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.251", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "2dd81bdb01ff4c39a4959f4af2acf61a", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "fb143ef7-8271-4a8a-a4aa-8eba9a89f6a1", "external-id": "nsx-vlan-transportzone-504", "segmentation_id": 504, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap8600276a-33", "ovs_interfaceid": "8600276a-3346-44f1-a760-df6caca2e51e", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1193.391755] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1193.534070] env[62753]: DEBUG nova.compute.manager [req-aa6d0b1f-afcc-4d07-9a97-82dd5738e012 req-c0828fe9-02d6-428c-8372-2dd2dc9f0ab4 service nova] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Received event network-vif-deleted-f7c88469-8383-42ae-a124-f4fdadc03fa2 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1193.749357] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1193.749581] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.175s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1193.749787] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.358s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1193.750019] env[62753]: DEBUG nova.objects.instance [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lazy-loading 'resources' on Instance uuid 21e6cf03-e528-4289-8813-86020a4e8d2a {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1193.848849] env[62753]: DEBUG oslo_concurrency.lockutils [req-ca06d230-ca6c-4caf-8d86-cdc76211e3c1 req-b200b589-7de3-4718-a28c-947c3151fcc3 service nova] Releasing lock "refresh_cache-ec0b07b8-1cb9-4911-8b03-efb466a21888" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1194.331768] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-de3baa92-a6bf-41ce-8a41-d4967d9ca938 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.339647] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-837fcebe-71c8-413c-86c5-dfcb03308101 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.368469] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d8badcd0-2604-41c5-9680-aad56754b844 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.375468] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c6a1450-332c-48e1-90b3-c953bf0e345d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1194.388110] env[62753]: DEBUG nova.compute.provider_tree [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1194.890911] env[62753]: DEBUG nova.scheduler.client.report [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1195.398544] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.649s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1195.419477] env[62753]: INFO nova.scheduler.client.report [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Deleted allocations for instance 21e6cf03-e528-4289-8813-86020a4e8d2a [ 1195.752035] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1195.927677] env[62753]: DEBUG oslo_concurrency.lockutils [None req-d78f7d6a-6e56-4358-a9a6-74cb34b012c7 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "21e6cf03-e528-4289-8813-86020a4e8d2a" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 6.092s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1196.010797] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1196.010965] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Starting heal instance info cache {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1196.555489] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1196.555761] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquired lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1196.555802] env[62753]: DEBUG nova.network.neutron [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Forcefully refreshing network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1197.765967] env[62753]: DEBUG nova.network.neutron [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance_info_cache with network_info: [{"id": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "address": "fa:16:3e:88:6a:1d", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25e4c4c8-9d", "ovs_interfaceid": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1197.849970] env[62753]: DEBUG oslo_concurrency.lockutils [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "221f76f0-ee69-4119-b19b-c148e40c1126" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1197.850221] env[62753]: DEBUG oslo_concurrency.lockutils [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "221f76f0-ee69-4119-b19b-c148e40c1126" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1198.268472] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Releasing lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1198.268705] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updated the network info_cache for instance {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1198.352543] env[62753]: DEBUG nova.compute.manager [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1198.873164] env[62753]: DEBUG oslo_concurrency.lockutils [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1198.873459] env[62753]: DEBUG oslo_concurrency.lockutils [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1198.875029] env[62753]: INFO nova.compute.claims [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1199.963052] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-512f4568-196b-4244-a670-78774c84ab26 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1199.970526] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ff68a5bd-003f-48b6-8f5b-c6b3a3fdd400 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.001290] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-53dd6c78-dc0f-476b-a971-84d62241ab22 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.008476] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-367c74bc-eba3-4e72-b2c1-7de48cde91f8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1200.021164] env[62753]: DEBUG nova.compute.provider_tree [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1200.524311] env[62753]: DEBUG nova.scheduler.client.report [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1201.029410] env[62753]: DEBUG oslo_concurrency.lockutils [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.156s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1201.030047] env[62753]: DEBUG nova.compute.manager [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1201.535566] env[62753]: DEBUG nova.compute.utils [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1201.536925] env[62753]: DEBUG nova.compute.manager [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1201.537112] env[62753]: DEBUG nova.network.neutron [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1201.583537] env[62753]: DEBUG nova.policy [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': '592d1d366cc4461299dbc28cee63e5b4', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '8925b84dcf9a47fbaf2eb8044b3850fa', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1201.848195] env[62753]: DEBUG nova.network.neutron [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Successfully created port: 7d3a7c1e-e1c1-42da-aacb-dad06fa6049a {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1202.041103] env[62753]: DEBUG nova.compute.manager [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1202.990776] env[62753]: DEBUG oslo_concurrency.lockutils [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "40c28133-7363-4a8e-bd77-fce3bbfbee69" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1202.991015] env[62753]: DEBUG oslo_concurrency.lockutils [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "40c28133-7363-4a8e-bd77-fce3bbfbee69" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.051917] env[62753]: DEBUG nova.compute.manager [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1203.079929] env[62753]: DEBUG nova.virt.hardware [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1203.080186] env[62753]: DEBUG nova.virt.hardware [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1203.080375] env[62753]: DEBUG nova.virt.hardware [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1203.080588] env[62753]: DEBUG nova.virt.hardware [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1203.080742] env[62753]: DEBUG nova.virt.hardware [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1203.080898] env[62753]: DEBUG nova.virt.hardware [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1203.081118] env[62753]: DEBUG nova.virt.hardware [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1203.081295] env[62753]: DEBUG nova.virt.hardware [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1203.081459] env[62753]: DEBUG nova.virt.hardware [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1203.081623] env[62753]: DEBUG nova.virt.hardware [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1203.081799] env[62753]: DEBUG nova.virt.hardware [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1203.082706] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a80c3ab-3476-4361-af4c-9141cf0c5914 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.090490] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-efe926ff-70bb-445f-8596-5051560c2b91 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1203.274189] env[62753]: DEBUG nova.compute.manager [req-ea7e36cb-2c6c-4816-87b9-d2cb17d2cea7 req-f676edc7-0eec-4bf2-a594-eebcbbf97bf8 service nova] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Received event network-vif-plugged-7d3a7c1e-e1c1-42da-aacb-dad06fa6049a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1203.274408] env[62753]: DEBUG oslo_concurrency.lockutils [req-ea7e36cb-2c6c-4816-87b9-d2cb17d2cea7 req-f676edc7-0eec-4bf2-a594-eebcbbf97bf8 service nova] Acquiring lock "221f76f0-ee69-4119-b19b-c148e40c1126-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1203.274628] env[62753]: DEBUG oslo_concurrency.lockutils [req-ea7e36cb-2c6c-4816-87b9-d2cb17d2cea7 req-f676edc7-0eec-4bf2-a594-eebcbbf97bf8 service nova] Lock "221f76f0-ee69-4119-b19b-c148e40c1126-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1203.274798] env[62753]: DEBUG oslo_concurrency.lockutils [req-ea7e36cb-2c6c-4816-87b9-d2cb17d2cea7 req-f676edc7-0eec-4bf2-a594-eebcbbf97bf8 service nova] Lock "221f76f0-ee69-4119-b19b-c148e40c1126-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1203.275029] env[62753]: DEBUG nova.compute.manager [req-ea7e36cb-2c6c-4816-87b9-d2cb17d2cea7 req-f676edc7-0eec-4bf2-a594-eebcbbf97bf8 service nova] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] No waiting events found dispatching network-vif-plugged-7d3a7c1e-e1c1-42da-aacb-dad06fa6049a {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1203.275298] env[62753]: WARNING nova.compute.manager [req-ea7e36cb-2c6c-4816-87b9-d2cb17d2cea7 req-f676edc7-0eec-4bf2-a594-eebcbbf97bf8 service nova] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Received unexpected event network-vif-plugged-7d3a7c1e-e1c1-42da-aacb-dad06fa6049a for instance with vm_state building and task_state spawning. [ 1203.364999] env[62753]: DEBUG nova.network.neutron [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Successfully updated port: 7d3a7c1e-e1c1-42da-aacb-dad06fa6049a {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1203.493169] env[62753]: DEBUG nova.compute.manager [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1203.867775] env[62753]: DEBUG oslo_concurrency.lockutils [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "refresh_cache-221f76f0-ee69-4119-b19b-c148e40c1126" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1203.867943] env[62753]: DEBUG oslo_concurrency.lockutils [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquired lock "refresh_cache-221f76f0-ee69-4119-b19b-c148e40c1126" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1203.868166] env[62753]: DEBUG nova.network.neutron [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1204.016148] env[62753]: DEBUG oslo_concurrency.lockutils [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1204.016443] env[62753]: DEBUG oslo_concurrency.lockutils [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1204.018282] env[62753]: INFO nova.compute.claims [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1204.402390] env[62753]: DEBUG nova.network.neutron [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1204.563978] env[62753]: DEBUG nova.network.neutron [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Updating instance_info_cache with network_info: [{"id": "7d3a7c1e-e1c1-42da-aacb-dad06fa6049a", "address": "fa:16:3e:65:8d:11", "network": {"id": "61c6e640-d2e6-4b3b-b191-7da9249dd339", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-703836479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8925b84dcf9a47fbaf2eb8044b3850fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d3a7c1e-e1", "ovs_interfaceid": "7d3a7c1e-e1c1-42da-aacb-dad06fa6049a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1205.066504] env[62753]: DEBUG oslo_concurrency.lockutils [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Releasing lock "refresh_cache-221f76f0-ee69-4119-b19b-c148e40c1126" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1205.066817] env[62753]: DEBUG nova.compute.manager [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Instance network_info: |[{"id": "7d3a7c1e-e1c1-42da-aacb-dad06fa6049a", "address": "fa:16:3e:65:8d:11", "network": {"id": "61c6e640-d2e6-4b3b-b191-7da9249dd339", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-703836479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8925b84dcf9a47fbaf2eb8044b3850fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d3a7c1e-e1", "ovs_interfaceid": "7d3a7c1e-e1c1-42da-aacb-dad06fa6049a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1205.067265] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:65:8d:11', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': '510d3c47-3615-43d5-aa5d-a279fd915e71', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '7d3a7c1e-e1c1-42da-aacb-dad06fa6049a', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1205.074769] env[62753]: DEBUG oslo.service.loopingcall [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1205.077351] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1205.077781] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f92e7e4b-36f5-488c-9c3e-0c957686eb06 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.108910] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1205.108910] env[62753]: value = "task-1332936" [ 1205.108910] env[62753]: _type = "Task" [ 1205.108910] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.121936] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332936, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.149059] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9fa4b6c5-5f88-496d-9edb-53c27991bee9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.155927] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8d044d9-e36d-497b-95ab-cc6646b1cf4d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.188065] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2ac5a72b-3121-4e48-a26a-8e024f6b504b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.195466] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df17e8f1-0479-48b9-b040-1145f478dbb2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.208510] env[62753]: DEBUG nova.compute.provider_tree [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1205.303957] env[62753]: DEBUG nova.compute.manager [req-08e71418-31dc-4622-84c9-c80d2ca6b31c req-0650ef03-2b70-4219-ad47-15b76098fba7 service nova] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Received event network-changed-7d3a7c1e-e1c1-42da-aacb-dad06fa6049a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1205.304225] env[62753]: DEBUG nova.compute.manager [req-08e71418-31dc-4622-84c9-c80d2ca6b31c req-0650ef03-2b70-4219-ad47-15b76098fba7 service nova] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Refreshing instance network info cache due to event network-changed-7d3a7c1e-e1c1-42da-aacb-dad06fa6049a. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1205.304597] env[62753]: DEBUG oslo_concurrency.lockutils [req-08e71418-31dc-4622-84c9-c80d2ca6b31c req-0650ef03-2b70-4219-ad47-15b76098fba7 service nova] Acquiring lock "refresh_cache-221f76f0-ee69-4119-b19b-c148e40c1126" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1205.304770] env[62753]: DEBUG oslo_concurrency.lockutils [req-08e71418-31dc-4622-84c9-c80d2ca6b31c req-0650ef03-2b70-4219-ad47-15b76098fba7 service nova] Acquired lock "refresh_cache-221f76f0-ee69-4119-b19b-c148e40c1126" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1205.304951] env[62753]: DEBUG nova.network.neutron [req-08e71418-31dc-4622-84c9-c80d2ca6b31c req-0650ef03-2b70-4219-ad47-15b76098fba7 service nova] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Refreshing network info cache for port 7d3a7c1e-e1c1-42da-aacb-dad06fa6049a {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1205.619198] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332936, 'name': CreateVM_Task, 'duration_secs': 0.302604} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1205.620109] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1205.620109] env[62753]: DEBUG oslo_concurrency.lockutils [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1205.620225] env[62753]: DEBUG oslo_concurrency.lockutils [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1205.620511] env[62753]: DEBUG oslo_concurrency.lockutils [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1205.620759] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-fc5f9c17-ca9c-40e7-9a18-187ae8dd5977 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1205.625013] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1205.625013] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52532f3e-67db-44a5-5534-68a846e129a2" [ 1205.625013] env[62753]: _type = "Task" [ 1205.625013] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1205.632778] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52532f3e-67db-44a5-5534-68a846e129a2, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1205.711983] env[62753]: DEBUG nova.scheduler.client.report [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1206.002014] env[62753]: DEBUG nova.network.neutron [req-08e71418-31dc-4622-84c9-c80d2ca6b31c req-0650ef03-2b70-4219-ad47-15b76098fba7 service nova] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Updated VIF entry in instance network info cache for port 7d3a7c1e-e1c1-42da-aacb-dad06fa6049a. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1206.002408] env[62753]: DEBUG nova.network.neutron [req-08e71418-31dc-4622-84c9-c80d2ca6b31c req-0650ef03-2b70-4219-ad47-15b76098fba7 service nova] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Updating instance_info_cache with network_info: [{"id": "7d3a7c1e-e1c1-42da-aacb-dad06fa6049a", "address": "fa:16:3e:65:8d:11", "network": {"id": "61c6e640-d2e6-4b3b-b191-7da9249dd339", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-703836479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8925b84dcf9a47fbaf2eb8044b3850fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d3a7c1e-e1", "ovs_interfaceid": "7d3a7c1e-e1c1-42da-aacb-dad06fa6049a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1206.136356] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52532f3e-67db-44a5-5534-68a846e129a2, 'name': SearchDatastore_Task, 'duration_secs': 0.012866} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.136603] env[62753]: DEBUG oslo_concurrency.lockutils [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1206.136844] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1206.137104] env[62753]: DEBUG oslo_concurrency.lockutils [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1206.137259] env[62753]: DEBUG oslo_concurrency.lockutils [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1206.137445] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1206.137709] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-f3a84c9d-0b7f-45af-b6ca-f6f508b7a3a3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.146954] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1206.147142] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1206.147838] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8ff757d4-0458-4d12-a3f0-536fd77b5cd5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.152795] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1206.152795] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bc94e7-137f-d785-b402-2093b40a0564" [ 1206.152795] env[62753]: _type = "Task" [ 1206.152795] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.160158] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bc94e7-137f-d785-b402-2093b40a0564, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.217203] env[62753]: DEBUG oslo_concurrency.lockutils [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.201s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1206.217837] env[62753]: DEBUG nova.compute.manager [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1206.505593] env[62753]: DEBUG oslo_concurrency.lockutils [req-08e71418-31dc-4622-84c9-c80d2ca6b31c req-0650ef03-2b70-4219-ad47-15b76098fba7 service nova] Releasing lock "refresh_cache-221f76f0-ee69-4119-b19b-c148e40c1126" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1206.663102] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bc94e7-137f-d785-b402-2093b40a0564, 'name': SearchDatastore_Task, 'duration_secs': 0.008908} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1206.663873] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-0e5b102c-96f0-4dc0-872a-2f89d2b5d2f9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1206.668741] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1206.668741] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]526f8a58-4872-aa86-ffe4-799547256968" [ 1206.668741] env[62753]: _type = "Task" [ 1206.668741] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1206.676025] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]526f8a58-4872-aa86-ffe4-799547256968, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1206.722349] env[62753]: DEBUG nova.compute.utils [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1206.723616] env[62753]: DEBUG nova.compute.manager [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1206.723784] env[62753]: DEBUG nova.network.neutron [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1206.760365] env[62753]: DEBUG nova.policy [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd65bf0df728640ef9104e14d6320ed01', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '891d1efbcac34ceba5fa72ea6b5fe38b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1207.013698] env[62753]: DEBUG nova.network.neutron [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Successfully created port: 12b306fe-5e83-4400-b6f9-c68592fb9cca {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1207.179378] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]526f8a58-4872-aa86-ffe4-799547256968, 'name': SearchDatastore_Task, 'duration_secs': 0.009381} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.179643] env[62753]: DEBUG oslo_concurrency.lockutils [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1207.179905] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 221f76f0-ee69-4119-b19b-c148e40c1126/221f76f0-ee69-4119-b19b-c148e40c1126.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1207.180419] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-980d144e-8bcd-42a2-b398-7de61d806b4a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.186460] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1207.186460] env[62753]: value = "task-1332937" [ 1207.186460] env[62753]: _type = "Task" [ 1207.186460] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.193560] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332937, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1207.226412] env[62753]: DEBUG nova.compute.manager [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1207.695667] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332937, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.487657} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1207.696926] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 221f76f0-ee69-4119-b19b-c148e40c1126/221f76f0-ee69-4119-b19b-c148e40c1126.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1207.696926] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1207.697093] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-31424ebe-c467-46ad-ad26-6d255c0f4e09 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1207.704066] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1207.704066] env[62753]: value = "task-1332938" [ 1207.704066] env[62753]: _type = "Task" [ 1207.704066] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1207.710744] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332938, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.213528] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332938, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.06511} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.213528] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1208.214312] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3bb44af4-721c-4fe3-8f07-d89b947e0277 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.235987] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Reconfiguring VM instance instance-0000006f to attach disk [datastore1] 221f76f0-ee69-4119-b19b-c148e40c1126/221f76f0-ee69-4119-b19b-c148e40c1126.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1208.237092] env[62753]: DEBUG nova.compute.manager [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1208.238941] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cbd7207f-ddf3-4fe2-8813-feb6bdacff5c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.260895] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1208.260895] env[62753]: value = "task-1332939" [ 1208.260895] env[62753]: _type = "Task" [ 1208.260895] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.269571] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332939, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1208.277067] env[62753]: DEBUG nova.virt.hardware [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1208.277332] env[62753]: DEBUG nova.virt.hardware [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1208.277511] env[62753]: DEBUG nova.virt.hardware [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1208.277700] env[62753]: DEBUG nova.virt.hardware [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1208.277852] env[62753]: DEBUG nova.virt.hardware [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1208.278012] env[62753]: DEBUG nova.virt.hardware [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1208.278236] env[62753]: DEBUG nova.virt.hardware [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1208.278399] env[62753]: DEBUG nova.virt.hardware [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1208.278572] env[62753]: DEBUG nova.virt.hardware [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1208.278743] env[62753]: DEBUG nova.virt.hardware [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1208.278966] env[62753]: DEBUG nova.virt.hardware [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1208.280089] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ddb59d6f-c870-4f0d-b034-7295207806d1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.287326] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-56b586a0-3f87-47e5-95b5-afdd1d3ed717 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.577201] env[62753]: DEBUG nova.compute.manager [req-0c967240-b73e-47ce-a9fa-057e304c2b8d req-a43830b4-d940-481b-985e-a91e7330b9b5 service nova] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Received event network-vif-plugged-12b306fe-5e83-4400-b6f9-c68592fb9cca {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1208.577404] env[62753]: DEBUG oslo_concurrency.lockutils [req-0c967240-b73e-47ce-a9fa-057e304c2b8d req-a43830b4-d940-481b-985e-a91e7330b9b5 service nova] Acquiring lock "40c28133-7363-4a8e-bd77-fce3bbfbee69-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1208.577629] env[62753]: DEBUG oslo_concurrency.lockutils [req-0c967240-b73e-47ce-a9fa-057e304c2b8d req-a43830b4-d940-481b-985e-a91e7330b9b5 service nova] Lock "40c28133-7363-4a8e-bd77-fce3bbfbee69-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1208.577801] env[62753]: DEBUG oslo_concurrency.lockutils [req-0c967240-b73e-47ce-a9fa-057e304c2b8d req-a43830b4-d940-481b-985e-a91e7330b9b5 service nova] Lock "40c28133-7363-4a8e-bd77-fce3bbfbee69-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1208.577968] env[62753]: DEBUG nova.compute.manager [req-0c967240-b73e-47ce-a9fa-057e304c2b8d req-a43830b4-d940-481b-985e-a91e7330b9b5 service nova] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] No waiting events found dispatching network-vif-plugged-12b306fe-5e83-4400-b6f9-c68592fb9cca {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1208.578187] env[62753]: WARNING nova.compute.manager [req-0c967240-b73e-47ce-a9fa-057e304c2b8d req-a43830b4-d940-481b-985e-a91e7330b9b5 service nova] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Received unexpected event network-vif-plugged-12b306fe-5e83-4400-b6f9-c68592fb9cca for instance with vm_state building and task_state spawning. [ 1208.658546] env[62753]: DEBUG nova.network.neutron [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Successfully updated port: 12b306fe-5e83-4400-b6f9-c68592fb9cca {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1208.771019] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332939, 'name': ReconfigVM_Task, 'duration_secs': 0.285953} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1208.771363] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Reconfigured VM instance instance-0000006f to attach disk [datastore1] 221f76f0-ee69-4119-b19b-c148e40c1126/221f76f0-ee69-4119-b19b-c148e40c1126.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1208.771955] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-2816e57e-ab05-4adb-8719-72728fbf51ab {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1208.778624] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1208.778624] env[62753]: value = "task-1332940" [ 1208.778624] env[62753]: _type = "Task" [ 1208.778624] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1208.785854] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332940, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.162048] env[62753]: DEBUG oslo_concurrency.lockutils [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "refresh_cache-40c28133-7363-4a8e-bd77-fce3bbfbee69" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1209.162048] env[62753]: DEBUG oslo_concurrency.lockutils [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "refresh_cache-40c28133-7363-4a8e-bd77-fce3bbfbee69" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1209.162048] env[62753]: DEBUG nova.network.neutron [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1209.287821] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332940, 'name': Rename_Task, 'duration_secs': 0.140618} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.288108] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1209.288343] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-f98853b4-259a-4a4b-946a-11b317831f1f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.296396] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1209.296396] env[62753]: value = "task-1332941" [ 1209.296396] env[62753]: _type = "Task" [ 1209.296396] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1209.312566] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332941, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1209.694181] env[62753]: DEBUG nova.network.neutron [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1209.806062] env[62753]: DEBUG oslo_vmware.api [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332941, 'name': PowerOnVM_Task, 'duration_secs': 0.428092} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1209.806421] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1209.806602] env[62753]: INFO nova.compute.manager [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Took 6.75 seconds to spawn the instance on the hypervisor. [ 1209.806784] env[62753]: DEBUG nova.compute.manager [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1209.807574] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0d98e7d4-299e-4516-a981-c45e8cd1b0e8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1209.818810] env[62753]: DEBUG nova.network.neutron [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Updating instance_info_cache with network_info: [{"id": "12b306fe-5e83-4400-b6f9-c68592fb9cca", "address": "fa:16:3e:87:22:d1", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12b306fe-5e", "ovs_interfaceid": "12b306fe-5e83-4400-b6f9-c68592fb9cca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1210.321788] env[62753]: DEBUG oslo_concurrency.lockutils [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "refresh_cache-40c28133-7363-4a8e-bd77-fce3bbfbee69" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1210.321788] env[62753]: DEBUG nova.compute.manager [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Instance network_info: |[{"id": "12b306fe-5e83-4400-b6f9-c68592fb9cca", "address": "fa:16:3e:87:22:d1", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12b306fe-5e", "ovs_interfaceid": "12b306fe-5e83-4400-b6f9-c68592fb9cca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1210.323279] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:22:d1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cae1d6a8-cbba-4bbf-af10-ba5467340475', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '12b306fe-5e83-4400-b6f9-c68592fb9cca', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1210.331795] env[62753]: DEBUG oslo.service.loopingcall [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1210.332188] env[62753]: INFO nova.compute.manager [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Took 11.47 seconds to build instance. [ 1210.332971] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1210.333589] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-62390485-203d-49cf-ba4f-4be734c39c47 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.348327] env[62753]: DEBUG oslo_concurrency.lockutils [None req-829eac29-67cf-4a0c-9192-81f783488d6b tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "221f76f0-ee69-4119-b19b-c148e40c1126" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.498s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1210.355421] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1210.355421] env[62753]: value = "task-1332942" [ 1210.355421] env[62753]: _type = "Task" [ 1210.355421] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.365863] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332942, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1210.609984] env[62753]: DEBUG nova.compute.manager [req-980b7787-76ac-43cf-83b0-47bc1536ea00 req-e091bab0-203b-4dfb-8073-f73c5a1680c4 service nova] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Received event network-changed-12b306fe-5e83-4400-b6f9-c68592fb9cca {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1210.609984] env[62753]: DEBUG nova.compute.manager [req-980b7787-76ac-43cf-83b0-47bc1536ea00 req-e091bab0-203b-4dfb-8073-f73c5a1680c4 service nova] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Refreshing instance network info cache due to event network-changed-12b306fe-5e83-4400-b6f9-c68592fb9cca. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1210.609984] env[62753]: DEBUG oslo_concurrency.lockutils [req-980b7787-76ac-43cf-83b0-47bc1536ea00 req-e091bab0-203b-4dfb-8073-f73c5a1680c4 service nova] Acquiring lock "refresh_cache-40c28133-7363-4a8e-bd77-fce3bbfbee69" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1210.609984] env[62753]: DEBUG oslo_concurrency.lockutils [req-980b7787-76ac-43cf-83b0-47bc1536ea00 req-e091bab0-203b-4dfb-8073-f73c5a1680c4 service nova] Acquired lock "refresh_cache-40c28133-7363-4a8e-bd77-fce3bbfbee69" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1210.610402] env[62753]: DEBUG nova.network.neutron [req-980b7787-76ac-43cf-83b0-47bc1536ea00 req-e091bab0-203b-4dfb-8073-f73c5a1680c4 service nova] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Refreshing network info cache for port 12b306fe-5e83-4400-b6f9-c68592fb9cca {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1210.861984] env[62753]: DEBUG nova.compute.manager [req-47a191c3-75c1-443b-b4cb-6411d27c0f97 req-e46fa915-aa53-4d47-8dbd-749528f3a562 service nova] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Received event network-changed-7d3a7c1e-e1c1-42da-aacb-dad06fa6049a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1210.862231] env[62753]: DEBUG nova.compute.manager [req-47a191c3-75c1-443b-b4cb-6411d27c0f97 req-e46fa915-aa53-4d47-8dbd-749528f3a562 service nova] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Refreshing instance network info cache due to event network-changed-7d3a7c1e-e1c1-42da-aacb-dad06fa6049a. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1210.862411] env[62753]: DEBUG oslo_concurrency.lockutils [req-47a191c3-75c1-443b-b4cb-6411d27c0f97 req-e46fa915-aa53-4d47-8dbd-749528f3a562 service nova] Acquiring lock "refresh_cache-221f76f0-ee69-4119-b19b-c148e40c1126" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1210.862559] env[62753]: DEBUG oslo_concurrency.lockutils [req-47a191c3-75c1-443b-b4cb-6411d27c0f97 req-e46fa915-aa53-4d47-8dbd-749528f3a562 service nova] Acquired lock "refresh_cache-221f76f0-ee69-4119-b19b-c148e40c1126" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1210.862723] env[62753]: DEBUG nova.network.neutron [req-47a191c3-75c1-443b-b4cb-6411d27c0f97 req-e46fa915-aa53-4d47-8dbd-749528f3a562 service nova] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Refreshing network info cache for port 7d3a7c1e-e1c1-42da-aacb-dad06fa6049a {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1210.869432] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332942, 'name': CreateVM_Task, 'duration_secs': 0.303894} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1210.869615] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1210.870214] env[62753]: DEBUG oslo_concurrency.lockutils [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1210.870705] env[62753]: DEBUG oslo_concurrency.lockutils [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1210.870705] env[62753]: DEBUG oslo_concurrency.lockutils [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1210.871375] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-900c7eaf-4bc3-4bf2-8d65-6181aabf5187 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1210.876208] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1210.876208] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525dfd31-8d95-9ea0-8d95-7f68615612c7" [ 1210.876208] env[62753]: _type = "Task" [ 1210.876208] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1210.883907] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525dfd31-8d95-9ea0-8d95-7f68615612c7, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.321850] env[62753]: DEBUG nova.network.neutron [req-980b7787-76ac-43cf-83b0-47bc1536ea00 req-e091bab0-203b-4dfb-8073-f73c5a1680c4 service nova] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Updated VIF entry in instance network info cache for port 12b306fe-5e83-4400-b6f9-c68592fb9cca. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1211.322236] env[62753]: DEBUG nova.network.neutron [req-980b7787-76ac-43cf-83b0-47bc1536ea00 req-e091bab0-203b-4dfb-8073-f73c5a1680c4 service nova] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Updating instance_info_cache with network_info: [{"id": "12b306fe-5e83-4400-b6f9-c68592fb9cca", "address": "fa:16:3e:87:22:d1", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.5", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap12b306fe-5e", "ovs_interfaceid": "12b306fe-5e83-4400-b6f9-c68592fb9cca", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1211.386221] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]525dfd31-8d95-9ea0-8d95-7f68615612c7, 'name': SearchDatastore_Task, 'duration_secs': 0.010909} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.386552] env[62753]: DEBUG oslo_concurrency.lockutils [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1211.386796] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1211.387039] env[62753]: DEBUG oslo_concurrency.lockutils [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1211.387195] env[62753]: DEBUG oslo_concurrency.lockutils [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1211.387380] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1211.387642] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-95f435f7-6458-47ff-af5c-098bff6b3b02 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.396215] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1211.396419] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1211.397388] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c3110e4b-502b-4b86-b6f7-cdee938e920a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.402284] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1211.402284] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f86f38-d274-b5f2-2504-c7324e04a8ec" [ 1211.402284] env[62753]: _type = "Task" [ 1211.402284] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.409684] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f86f38-d274-b5f2-2504-c7324e04a8ec, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1211.567260] env[62753]: DEBUG nova.network.neutron [req-47a191c3-75c1-443b-b4cb-6411d27c0f97 req-e46fa915-aa53-4d47-8dbd-749528f3a562 service nova] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Updated VIF entry in instance network info cache for port 7d3a7c1e-e1c1-42da-aacb-dad06fa6049a. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1211.567641] env[62753]: DEBUG nova.network.neutron [req-47a191c3-75c1-443b-b4cb-6411d27c0f97 req-e46fa915-aa53-4d47-8dbd-749528f3a562 service nova] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Updating instance_info_cache with network_info: [{"id": "7d3a7c1e-e1c1-42da-aacb-dad06fa6049a", "address": "fa:16:3e:65:8d:11", "network": {"id": "61c6e640-d2e6-4b3b-b191-7da9249dd339", "bridge": "br-int", "label": "tempest-AttachVolumeNegativeTest-703836479-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.12", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.190", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "8925b84dcf9a47fbaf2eb8044b3850fa", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "510d3c47-3615-43d5-aa5d-a279fd915e71", "external-id": "nsx-vlan-transportzone-436", "segmentation_id": 436, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap7d3a7c1e-e1", "ovs_interfaceid": "7d3a7c1e-e1c1-42da-aacb-dad06fa6049a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1211.825231] env[62753]: DEBUG oslo_concurrency.lockutils [req-980b7787-76ac-43cf-83b0-47bc1536ea00 req-e091bab0-203b-4dfb-8073-f73c5a1680c4 service nova] Releasing lock "refresh_cache-40c28133-7363-4a8e-bd77-fce3bbfbee69" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1211.913488] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f86f38-d274-b5f2-2504-c7324e04a8ec, 'name': SearchDatastore_Task, 'duration_secs': 0.009041} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1211.914301] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-1ba477d1-e94c-4aea-adbd-cc8c5253f4c9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1211.919478] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1211.919478] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52113700-30be-f499-2b8c-7927e35df530" [ 1211.919478] env[62753]: _type = "Task" [ 1211.919478] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1211.926507] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52113700-30be-f499-2b8c-7927e35df530, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.069882] env[62753]: DEBUG oslo_concurrency.lockutils [req-47a191c3-75c1-443b-b4cb-6411d27c0f97 req-e46fa915-aa53-4d47-8dbd-749528f3a562 service nova] Releasing lock "refresh_cache-221f76f0-ee69-4119-b19b-c148e40c1126" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1212.430266] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52113700-30be-f499-2b8c-7927e35df530, 'name': SearchDatastore_Task, 'duration_secs': 0.008484} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.430605] env[62753]: DEBUG oslo_concurrency.lockutils [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1212.430833] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 40c28133-7363-4a8e-bd77-fce3bbfbee69/40c28133-7363-4a8e-bd77-fce3bbfbee69.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1212.431124] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-fae26b6e-b475-435a-aa10-e4f9bd6f5cf0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.437889] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1212.437889] env[62753]: value = "task-1332943" [ 1212.437889] env[62753]: _type = "Task" [ 1212.437889] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.447013] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332943, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1212.948186] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332943, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.459878} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1212.948477] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 40c28133-7363-4a8e-bd77-fce3bbfbee69/40c28133-7363-4a8e-bd77-fce3bbfbee69.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1212.948660] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1212.948929] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-642cf0d4-2ff1-48ee-80fc-09605a8b1a94 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1212.954884] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1212.954884] env[62753]: value = "task-1332944" [ 1212.954884] env[62753]: _type = "Task" [ 1212.954884] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1212.962454] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332944, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1213.463972] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332944, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.060473} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1213.465365] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1213.465365] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba08da2d-bf8e-452b-8bf3-b74c939a6404 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.486436] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] 40c28133-7363-4a8e-bd77-fce3bbfbee69/40c28133-7363-4a8e-bd77-fce3bbfbee69.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1213.486659] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-3cdd6c79-d8a0-4819-a3ec-dba82aec82ae {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1213.504781] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1213.504781] env[62753]: value = "task-1332945" [ 1213.504781] env[62753]: _type = "Task" [ 1213.504781] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1213.512052] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332945, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.015194] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332945, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1214.514950] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332945, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.015864] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332945, 'name': ReconfigVM_Task, 'duration_secs': 1.083844} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.016258] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Reconfigured VM instance instance-00000070 to attach disk [datastore1] 40c28133-7363-4a8e-bd77-fce3bbfbee69/40c28133-7363-4a8e-bd77-fce3bbfbee69.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1215.016785] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-ee5ab5b7-e0fa-4017-bc60-1164c5647b2f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.023338] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1215.023338] env[62753]: value = "task-1332946" [ 1215.023338] env[62753]: _type = "Task" [ 1215.023338] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.033210] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332946, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1215.532866] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332946, 'name': Rename_Task, 'duration_secs': 0.132597} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1215.533164] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1215.533409] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-0188f6f4-909d-4644-a805-ee9d53531424 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1215.539175] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1215.539175] env[62753]: value = "task-1332947" [ 1215.539175] env[62753]: _type = "Task" [ 1215.539175] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1215.546100] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332947, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.050373] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332947, 'name': PowerOnVM_Task} progress is 89%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1216.549542] env[62753]: DEBUG oslo_vmware.api [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332947, 'name': PowerOnVM_Task, 'duration_secs': 0.648147} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1216.549799] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1216.550008] env[62753]: INFO nova.compute.manager [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Took 8.31 seconds to spawn the instance on the hypervisor. [ 1216.550203] env[62753]: DEBUG nova.compute.manager [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1216.550930] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0a3ec72d-bfc0-4b25-8e87-822bfe241914 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1217.067651] env[62753]: INFO nova.compute.manager [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Took 13.07 seconds to build instance. [ 1217.532920] env[62753]: INFO nova.compute.manager [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Rebuilding instance [ 1217.570908] env[62753]: DEBUG oslo_concurrency.lockutils [None req-46b876ea-4892-4695-8c3f-ecb6f641fa1b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "40c28133-7363-4a8e-bd77-fce3bbfbee69" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.580s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1217.573686] env[62753]: DEBUG nova.compute.manager [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1217.574577] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-628d94e7-27da-4ab7-91df-f21188262f6b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.085444] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1218.085759] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-60410312-d89e-43f9-a7f9-c506e76dd377 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.092539] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1218.092539] env[62753]: value = "task-1332948" [ 1218.092539] env[62753]: _type = "Task" [ 1218.092539] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.100418] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332948, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1218.602852] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332948, 'name': PowerOffVM_Task, 'duration_secs': 0.195784} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1218.603138] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1218.603367] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1218.604110] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c5c2c74-9271-4239-a391-433dfcd8321d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.610250] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1218.610466] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-05f5baf0-c045-47e0-aec9-6021882e4c5a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.721462] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1218.721719] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1218.721921] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Deleting the datastore file [datastore1] 40c28133-7363-4a8e-bd77-fce3bbfbee69 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1218.722271] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b49bae3b-cba6-4af9-a9ff-1e6302a1687a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1218.729308] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1218.729308] env[62753]: value = "task-1332950" [ 1218.729308] env[62753]: _type = "Task" [ 1218.729308] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1218.736655] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332950, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1219.239249] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332950, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157151} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1219.239502] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1219.239682] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1219.239860] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1220.272745] env[62753]: DEBUG nova.virt.hardware [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1220.273047] env[62753]: DEBUG nova.virt.hardware [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1220.273178] env[62753]: DEBUG nova.virt.hardware [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1220.273367] env[62753]: DEBUG nova.virt.hardware [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1220.273520] env[62753]: DEBUG nova.virt.hardware [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1220.273674] env[62753]: DEBUG nova.virt.hardware [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1220.273887] env[62753]: DEBUG nova.virt.hardware [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1220.274069] env[62753]: DEBUG nova.virt.hardware [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1220.274252] env[62753]: DEBUG nova.virt.hardware [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1220.274435] env[62753]: DEBUG nova.virt.hardware [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1220.274642] env[62753]: DEBUG nova.virt.hardware [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1220.275521] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-666610e0-40ea-4a1c-9a37-4160361f4ac7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.283396] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2f563a82-2c55-4815-a02d-8ceb69f78988 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.297725] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:87:22:d1', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cae1d6a8-cbba-4bbf-af10-ba5467340475', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '12b306fe-5e83-4400-b6f9-c68592fb9cca', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1220.304883] env[62753]: DEBUG oslo.service.loopingcall [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1220.305116] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1220.305313] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d9485597-d152-42d0-94a6-537336ecd114 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.324085] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1220.324085] env[62753]: value = "task-1332951" [ 1220.324085] env[62753]: _type = "Task" [ 1220.324085] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.331131] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332951, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1220.833495] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332951, 'name': CreateVM_Task, 'duration_secs': 0.294657} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1220.833655] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1220.834369] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1220.834563] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1220.834893] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1220.835146] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-064e2c2e-d7b2-49e0-b0a3-f08450ed8fcf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1220.839345] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1220.839345] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a371e1-5075-3438-c0f0-cb6b106287db" [ 1220.839345] env[62753]: _type = "Task" [ 1220.839345] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1220.846198] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a371e1-5075-3438-c0f0-cb6b106287db, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.350013] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a371e1-5075-3438-c0f0-cb6b106287db, 'name': SearchDatastore_Task, 'duration_secs': 0.049202} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.350459] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1221.350655] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1221.350785] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1221.350937] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1221.351185] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1221.351490] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-7e7ea074-7127-4d20-8539-54c98aba1c5e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.359268] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1221.359432] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1221.360106] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-06d1bc67-c770-40e4-8daf-8430f32a5857 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.364519] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1221.364519] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5258de50-4bac-77e7-8646-9f11ab94b157" [ 1221.364519] env[62753]: _type = "Task" [ 1221.364519] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.371716] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5258de50-4bac-77e7-8646-9f11ab94b157, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1221.875418] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5258de50-4bac-77e7-8646-9f11ab94b157, 'name': SearchDatastore_Task, 'duration_secs': 0.007891} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1221.876160] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b1fb9b8c-5550-436b-ae0b-1a490151abbb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1221.881484] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1221.881484] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]526fd894-b64e-5c84-47f4-94a503c4a12f" [ 1221.881484] env[62753]: _type = "Task" [ 1221.881484] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1221.888820] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]526fd894-b64e-5c84-47f4-94a503c4a12f, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.392165] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]526fd894-b64e-5c84-47f4-94a503c4a12f, 'name': SearchDatastore_Task, 'duration_secs': 0.009428} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.392565] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1222.392688] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 40c28133-7363-4a8e-bd77-fce3bbfbee69/40c28133-7363-4a8e-bd77-fce3bbfbee69.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1222.392943] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-22c0664c-a3a4-4038-ba86-759dd5091c12 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.399803] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1222.399803] env[62753]: value = "task-1332952" [ 1222.399803] env[62753]: _type = "Task" [ 1222.399803] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.406861] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332952, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1222.909824] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332952, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.441531} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1222.910066] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 40c28133-7363-4a8e-bd77-fce3bbfbee69/40c28133-7363-4a8e-bd77-fce3bbfbee69.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1222.910289] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1222.910533] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-35fbd3d1-cd4d-4248-ae57-c06bb9834684 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1222.917131] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1222.917131] env[62753]: value = "task-1332953" [ 1222.917131] env[62753]: _type = "Task" [ 1222.917131] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1222.924723] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332953, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.426216] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332953, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.059076} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.426667] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1223.427285] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7aecd392-be61-40c6-a6b3-b9007b9892a1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.449206] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Reconfiguring VM instance instance-00000070 to attach disk [datastore1] 40c28133-7363-4a8e-bd77-fce3bbfbee69/40c28133-7363-4a8e-bd77-fce3bbfbee69.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1223.449442] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-c8446f52-b35f-4763-87b5-0fa8248ba653 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.468448] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1223.468448] env[62753]: value = "task-1332954" [ 1223.468448] env[62753]: _type = "Task" [ 1223.468448] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.477460] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332954, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1223.977805] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332954, 'name': ReconfigVM_Task, 'duration_secs': 0.287206} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1223.978093] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Reconfigured VM instance instance-00000070 to attach disk [datastore1] 40c28133-7363-4a8e-bd77-fce3bbfbee69/40c28133-7363-4a8e-bd77-fce3bbfbee69.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1223.978719] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-d25e45e1-8de6-4756-95af-67fe13cb1f9e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1223.985115] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1223.985115] env[62753]: value = "task-1332955" [ 1223.985115] env[62753]: _type = "Task" [ 1223.985115] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1223.992558] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332955, 'name': Rename_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.452275] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "2e784985-7045-448e-a650-f6692bb69d94" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1224.452654] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "2e784985-7045-448e-a650-f6692bb69d94" acquired by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1224.452727] env[62753]: INFO nova.compute.manager [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Shelving [ 1224.494748] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332955, 'name': Rename_Task, 'duration_secs': 0.333813} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1224.495018] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1224.495264] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-c328c8cf-bc32-4ec3-9abd-34090d0a92e1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.501471] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1224.501471] env[62753]: value = "task-1332956" [ 1224.501471] env[62753]: _type = "Task" [ 1224.501471] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.508659] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332956, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1224.960395] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1224.960732] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-ea2e8781-1ec7-43f1-a62d-ae4d1f6873bf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1224.967794] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1224.967794] env[62753]: value = "task-1332957" [ 1224.967794] env[62753]: _type = "Task" [ 1224.967794] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1224.976259] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332957, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1225.011110] env[62753]: DEBUG oslo_vmware.api [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332956, 'name': PowerOnVM_Task, 'duration_secs': 0.426894} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.011438] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1225.011706] env[62753]: DEBUG nova.compute.manager [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1225.012523] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-18cf00ec-7456-499c-ab0d-cc64209fd3a6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.477979] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332957, 'name': PowerOffVM_Task, 'duration_secs': 0.220284} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1225.478371] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1225.478977] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc7b7168-361f-474d-9392-2b539831f971 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.497237] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1dc5ba01-da2f-4b5a-a71e-b3bdfc63e145 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1225.528415] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1225.528415] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1225.528415] env[62753]: DEBUG nova.objects.instance [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Trying to apply a migration context that does not seem to be set for this instance {{(pid=62753) apply_migration_context /opt/stack/nova/nova/objects/instance.py:1067}} [ 1225.688658] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1225.688940] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1226.007892] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Creating Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:913}} [ 1226.008424] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CreateSnapshot_Task with opID=oslo.vmware-eecbd532-62fb-472f-8dff-c1639dcb3c7a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.017234] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1226.017234] env[62753]: value = "task-1332958" [ 1226.017234] env[62753]: _type = "Task" [ 1226.017234] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.025067] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332958, 'name': CreateSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.192611] env[62753]: INFO nova.compute.manager [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Detaching volume dc9f6473-9f05-4ca4-a314-4a1175d096bb [ 1226.223913] env[62753]: INFO nova.virt.block_device [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Attempting to driver detach volume dc9f6473-9f05-4ca4-a314-4a1175d096bb from mountpoint /dev/sdb [ 1226.223913] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Volume detach. Driver type: vmdk {{(pid=62753) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1226.223913] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284700', 'volume_id': 'dc9f6473-9f05-4ca4-a314-4a1175d096bb', 'name': 'volume-dc9f6473-9f05-4ca4-a314-4a1175d096bb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ec0b07b8-1cb9-4911-8b03-efb466a21888', 'attached_at': '', 'detached_at': '', 'volume_id': 'dc9f6473-9f05-4ca4-a314-4a1175d096bb', 'serial': 'dc9f6473-9f05-4ca4-a314-4a1175d096bb'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1226.225029] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1c2a46d8-629e-47a2-aa42-40fdbc8a0f1c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.245803] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0b8c106-adbe-4456-880c-7af8c694e396 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.252478] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21552c81-6a3c-469b-893a-c6da78945feb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.275181] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5f5e415a-fb3b-4292-a7c1-669679e9270e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.290830] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] The volume has not been displaced from its original location: [datastore1] volume-dc9f6473-9f05-4ca4-a314-4a1175d096bb/volume-dc9f6473-9f05-4ca4-a314-4a1175d096bb.vmdk. No consolidation needed. {{(pid=62753) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1226.297758] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Reconfiguring VM instance instance-0000006b to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1226.297985] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-752c921f-c9b6-403d-bcd3-0824ddd2bb0c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.317976] env[62753]: DEBUG oslo_vmware.api [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1226.317976] env[62753]: value = "task-1332959" [ 1226.317976] env[62753]: _type = "Task" [ 1226.317976] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.326162] env[62753]: DEBUG oslo_vmware.api [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332959, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.527746] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332958, 'name': CreateSnapshot_Task, 'duration_secs': 0.469264} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.528187] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Created Snapshot of the VM instance {{(pid=62753) _create_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:922}} [ 1226.528870] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4dce8e09-76f5-4cf5-a86e-15562df99946 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.536834] env[62753]: DEBUG oslo_concurrency.lockutils [None req-7fcec2e4-cc64-46ff-a543-2538db4160ef tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.finish_evacuation" :: held 1.009s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1226.831493] env[62753]: DEBUG oslo_vmware.api [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332959, 'name': ReconfigVM_Task, 'duration_secs': 0.232709} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1226.831762] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Reconfigured VM instance instance-0000006b to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1226.837595] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b27fa535-32a4-4527-8d33-0766f78eab55 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.851972] env[62753]: DEBUG oslo_vmware.api [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1226.851972] env[62753]: value = "task-1332960" [ 1226.851972] env[62753]: _type = "Task" [ 1226.851972] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.860229] env[62753]: DEBUG oslo_vmware.api [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332960, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1226.873701] env[62753]: DEBUG oslo_concurrency.lockutils [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "40c28133-7363-4a8e-bd77-fce3bbfbee69" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1226.873931] env[62753]: DEBUG oslo_concurrency.lockutils [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "40c28133-7363-4a8e-bd77-fce3bbfbee69" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1226.874158] env[62753]: DEBUG oslo_concurrency.lockutils [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "40c28133-7363-4a8e-bd77-fce3bbfbee69-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1226.874344] env[62753]: DEBUG oslo_concurrency.lockutils [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "40c28133-7363-4a8e-bd77-fce3bbfbee69-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1226.874554] env[62753]: DEBUG oslo_concurrency.lockutils [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "40c28133-7363-4a8e-bd77-fce3bbfbee69-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1226.876624] env[62753]: INFO nova.compute.manager [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Terminating instance [ 1226.878403] env[62753]: DEBUG nova.compute.manager [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1226.878597] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1226.879380] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6850f6e8-7dae-452d-a2e2-71ab254a0b10 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.885885] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1226.886118] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-e1a71113-38e3-40d6-a5d5-c38ff231dd8d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1226.892240] env[62753]: DEBUG oslo_vmware.api [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1226.892240] env[62753]: value = "task-1332961" [ 1226.892240] env[62753]: _type = "Task" [ 1226.892240] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1226.899451] env[62753]: DEBUG oslo_vmware.api [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332961, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.048187] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Creating linked-clone VM from snapshot {{(pid=62753) _create_linked_clone_from_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:955}} [ 1227.048479] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.CloneVM_Task with opID=oslo.vmware-7a8cb4f1-4eb1-4aee-a637-bce58fc3a814 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.057112] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1227.057112] env[62753]: value = "task-1332962" [ 1227.057112] env[62753]: _type = "Task" [ 1227.057112] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.065360] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332962, 'name': CloneVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.362963] env[62753]: DEBUG oslo_vmware.api [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332960, 'name': ReconfigVM_Task, 'duration_secs': 0.150921} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.363286] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284700', 'volume_id': 'dc9f6473-9f05-4ca4-a314-4a1175d096bb', 'name': 'volume-dc9f6473-9f05-4ca4-a314-4a1175d096bb', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': 'ec0b07b8-1cb9-4911-8b03-efb466a21888', 'attached_at': '', 'detached_at': '', 'volume_id': 'dc9f6473-9f05-4ca4-a314-4a1175d096bb', 'serial': 'dc9f6473-9f05-4ca4-a314-4a1175d096bb'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1227.402109] env[62753]: DEBUG oslo_vmware.api [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332961, 'name': PowerOffVM_Task, 'duration_secs': 0.20163} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1227.402739] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1227.402917] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1227.403192] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-b0145009-f1ac-4437-bc03-2ef88827baaa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.567325] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332962, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.750843] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1227.751130] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1227.751352] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Deleting the datastore file [datastore1] 40c28133-7363-4a8e-bd77-fce3bbfbee69 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1227.751709] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a74411cf-f9f4-4456-86a7-aa1d13f49ff3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1227.758060] env[62753]: DEBUG oslo_vmware.api [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1227.758060] env[62753]: value = "task-1332964" [ 1227.758060] env[62753]: _type = "Task" [ 1227.758060] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1227.766207] env[62753]: DEBUG oslo_vmware.api [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332964, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1227.904158] env[62753]: DEBUG nova.objects.instance [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lazy-loading 'flavor' on Instance uuid ec0b07b8-1cb9-4911-8b03-efb466a21888 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1228.068361] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332962, 'name': CloneVM_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1228.268154] env[62753]: DEBUG oslo_vmware.api [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332964, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.147121} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.268409] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1228.268603] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1228.268782] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1228.268959] env[62753]: INFO nova.compute.manager [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Took 1.39 seconds to destroy the instance on the hypervisor. [ 1228.269230] env[62753]: DEBUG oslo.service.loopingcall [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1228.269426] env[62753]: DEBUG nova.compute.manager [-] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1228.269521] env[62753]: DEBUG nova.network.neutron [-] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1228.549600] env[62753]: DEBUG nova.compute.manager [req-4e26cdbf-d79f-43a3-bc9e-c71cb1753d3c req-25f434da-91a2-456e-a816-f7a9cdf3aaf9 service nova] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Received event network-vif-deleted-12b306fe-5e83-4400-b6f9-c68592fb9cca {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1228.549787] env[62753]: INFO nova.compute.manager [req-4e26cdbf-d79f-43a3-bc9e-c71cb1753d3c req-25f434da-91a2-456e-a816-f7a9cdf3aaf9 service nova] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Neutron deleted interface 12b306fe-5e83-4400-b6f9-c68592fb9cca; detaching it from the instance and deleting it from the info cache [ 1228.549963] env[62753]: DEBUG nova.network.neutron [req-4e26cdbf-d79f-43a3-bc9e-c71cb1753d3c req-25f434da-91a2-456e-a816-f7a9cdf3aaf9 service nova] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1228.569809] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332962, 'name': CloneVM_Task, 'duration_secs': 1.452969} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1228.570092] env[62753]: INFO nova.virt.vmwareapi.vmops [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Created linked-clone VM from snapshot [ 1228.570765] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-807137a3-32c6-4ab7-bf66-130a83c2c2ee {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.578099] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Uploading image 026f5cd9-1845-4b09-9364-71568c7d3e7f {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:441}} [ 1228.601565] env[62753]: DEBUG oslo_vmware.rw_handles [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Creating HttpNfcLease lease for exporting VM: (result){ [ 1228.601565] env[62753]: value = "vm-284707" [ 1228.601565] env[62753]: _type = "VirtualMachine" [ 1228.601565] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:478}} [ 1228.601823] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ExportVm with opID=oslo.vmware-571e01ab-d36a-41b5-88f1-478333623fd6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1228.609214] env[62753]: DEBUG oslo_vmware.rw_handles [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lease: (returnval){ [ 1228.609214] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5269d190-9b2a-4af2-64a3-cf270f15d1da" [ 1228.609214] env[62753]: _type = "HttpNfcLease" [ 1228.609214] env[62753]: } obtained for exporting VM: (result){ [ 1228.609214] env[62753]: value = "vm-284707" [ 1228.609214] env[62753]: _type = "VirtualMachine" [ 1228.609214] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:481}} [ 1228.609493] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the lease: (returnval){ [ 1228.609493] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5269d190-9b2a-4af2-64a3-cf270f15d1da" [ 1228.609493] env[62753]: _type = "HttpNfcLease" [ 1228.609493] env[62753]: } to be ready. {{(pid=62753) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1228.615227] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1228.615227] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5269d190-9b2a-4af2-64a3-cf270f15d1da" [ 1228.615227] env[62753]: _type = "HttpNfcLease" [ 1228.615227] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1228.912039] env[62753]: DEBUG oslo_concurrency.lockutils [None req-cbc44c53-e400-4be6-88d2-0809a0bf3bc1 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.223s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1229.028137] env[62753]: DEBUG nova.network.neutron [-] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1229.054065] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-5b73e30b-9364-4030-af73-366981a10636 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.066039] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ce8243cd-98cf-4853-8386-0b9658b08442 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.091272] env[62753]: DEBUG nova.compute.manager [req-4e26cdbf-d79f-43a3-bc9e-c71cb1753d3c req-25f434da-91a2-456e-a816-f7a9cdf3aaf9 service nova] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Detach interface failed, port_id=12b306fe-5e83-4400-b6f9-c68592fb9cca, reason: Instance 40c28133-7363-4a8e-bd77-fce3bbfbee69 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1229.117150] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1229.117150] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5269d190-9b2a-4af2-64a3-cf270f15d1da" [ 1229.117150] env[62753]: _type = "HttpNfcLease" [ 1229.117150] env[62753]: } is ready. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1229.117446] env[62753]: DEBUG oslo_vmware.rw_handles [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1229.117446] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5269d190-9b2a-4af2-64a3-cf270f15d1da" [ 1229.117446] env[62753]: _type = "HttpNfcLease" [ 1229.117446] env[62753]: }. {{(pid=62753) _create_export_vm_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:486}} [ 1229.118176] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-00551e6b-8406-4619-9adb-4037b1cf85cf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.125189] env[62753]: DEBUG oslo_vmware.rw_handles [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52171f18-abe8-bca2-e849-c523d278cfdc/disk-0.vmdk from lease info. {{(pid=62753) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1229.125369] env[62753]: DEBUG oslo_vmware.rw_handles [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Opening URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52171f18-abe8-bca2-e849-c523d278cfdc/disk-0.vmdk for reading. {{(pid=62753) _create_read_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:99}} [ 1229.235582] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-319ed31c-de64-4b70-92a1-45837855b3f1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1229.530989] env[62753]: INFO nova.compute.manager [-] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Took 1.26 seconds to deallocate network for instance. [ 1230.039064] env[62753]: DEBUG oslo_concurrency.lockutils [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1230.039408] env[62753]: DEBUG oslo_concurrency.lockutils [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1230.039641] env[62753]: DEBUG nova.objects.instance [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lazy-loading 'resources' on Instance uuid 40c28133-7363-4a8e-bd77-fce3bbfbee69 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1230.096349] env[62753]: DEBUG oslo_concurrency.lockutils [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1230.096669] env[62753]: DEBUG oslo_concurrency.lockutils [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1230.096895] env[62753]: DEBUG oslo_concurrency.lockutils [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "ec0b07b8-1cb9-4911-8b03-efb466a21888-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1230.097103] env[62753]: DEBUG oslo_concurrency.lockutils [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "ec0b07b8-1cb9-4911-8b03-efb466a21888-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1230.097282] env[62753]: DEBUG oslo_concurrency.lockutils [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "ec0b07b8-1cb9-4911-8b03-efb466a21888-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1230.099538] env[62753]: INFO nova.compute.manager [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Terminating instance [ 1230.101432] env[62753]: DEBUG nova.compute.manager [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1230.101695] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1230.102602] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-30462851-87e0-4e81-8a7d-806fc6d57b85 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.110639] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1230.110940] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c9619e4e-e930-4ffb-9527-7fd0b123a8c6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.117946] env[62753]: DEBUG oslo_vmware.api [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1230.117946] env[62753]: value = "task-1332966" [ 1230.117946] env[62753]: _type = "Task" [ 1230.117946] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.126284] env[62753]: DEBUG oslo_vmware.api [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332966, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1230.628193] env[62753]: DEBUG oslo_vmware.api [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332966, 'name': PowerOffVM_Task, 'duration_secs': 0.182145} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1230.628582] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1230.628767] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1230.629136] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-df09d59c-40b2-4643-a002-a1f453c2c76d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.637093] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fcaf1eb3-3de1-44ad-9b6a-721ded4df5bd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.643741] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fed76216-5fcf-4523-b543-9fcc5a382935 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.674980] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-96258893-4898-41fb-ac2e-058730a413e2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.682626] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-194b59b4-94f3-4300-96fb-7919c61ed9db {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.696210] env[62753]: DEBUG nova.compute.provider_tree [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1230.699143] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1230.699385] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1230.699574] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Deleting the datastore file [datastore2] ec0b07b8-1cb9-4911-8b03-efb466a21888 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1230.700191] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-4547ab88-d35d-4cc7-9fb9-5be74ede5e18 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1230.705592] env[62753]: DEBUG oslo_vmware.api [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1230.705592] env[62753]: value = "task-1332968" [ 1230.705592] env[62753]: _type = "Task" [ 1230.705592] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1230.714612] env[62753]: DEBUG oslo_vmware.api [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332968, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1231.201105] env[62753]: DEBUG nova.scheduler.client.report [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1231.214924] env[62753]: DEBUG oslo_vmware.api [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332968, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.300811} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1231.215503] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1231.215716] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1231.215899] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1231.216087] env[62753]: INFO nova.compute.manager [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1231.216332] env[62753]: DEBUG oslo.service.loopingcall [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1231.217254] env[62753]: DEBUG nova.compute.manager [-] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1231.217353] env[62753]: DEBUG nova.network.neutron [-] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1231.671131] env[62753]: DEBUG nova.compute.manager [req-47cebe24-9c87-421a-8c3a-171274a3f210 req-c51203c1-ca63-45bb-aaf0-d7973f6f74e5 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Received event network-vif-deleted-8600276a-3346-44f1-a760-df6caca2e51e {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1231.671347] env[62753]: INFO nova.compute.manager [req-47cebe24-9c87-421a-8c3a-171274a3f210 req-c51203c1-ca63-45bb-aaf0-d7973f6f74e5 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Neutron deleted interface 8600276a-3346-44f1-a760-df6caca2e51e; detaching it from the instance and deleting it from the info cache [ 1231.671521] env[62753]: DEBUG nova.network.neutron [req-47cebe24-9c87-421a-8c3a-171274a3f210 req-c51203c1-ca63-45bb-aaf0-d7973f6f74e5 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1231.706675] env[62753]: DEBUG oslo_concurrency.lockutils [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.667s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1231.728345] env[62753]: INFO nova.scheduler.client.report [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Deleted allocations for instance 40c28133-7363-4a8e-bd77-fce3bbfbee69 [ 1232.149578] env[62753]: DEBUG nova.network.neutron [-] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1232.174270] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-9e5ef1da-536d-43d0-b006-967453a070e3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.187089] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9d1b5c84-aa3e-4781-bfa5-be9938270933 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1232.213857] env[62753]: DEBUG nova.compute.manager [req-47cebe24-9c87-421a-8c3a-171274a3f210 req-c51203c1-ca63-45bb-aaf0-d7973f6f74e5 service nova] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Detach interface failed, port_id=8600276a-3346-44f1-a760-df6caca2e51e, reason: Instance ec0b07b8-1cb9-4911-8b03-efb466a21888 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1232.236391] env[62753]: DEBUG oslo_concurrency.lockutils [None req-247746f2-4f87-4ab1-b42b-67cf80350053 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "40c28133-7363-4a8e-bd77-fce3bbfbee69" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.362s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1232.652587] env[62753]: INFO nova.compute.manager [-] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Took 1.43 seconds to deallocate network for instance. [ 1233.159734] env[62753]: DEBUG oslo_concurrency.lockutils [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1233.159984] env[62753]: DEBUG oslo_concurrency.lockutils [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1233.160239] env[62753]: DEBUG nova.objects.instance [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lazy-loading 'resources' on Instance uuid ec0b07b8-1cb9-4911-8b03-efb466a21888 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1233.552360] env[62753]: DEBUG nova.compute.manager [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Stashing vm_state: active {{(pid=62753) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1233.764646] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e9673bf3-f00c-4ff9-9bb6-768781630ae0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.773298] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-df7ea36c-d2bf-4f3d-9763-707f0ee1747d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.805576] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cd5a4d87-d325-4e81-99a5-6b12d28b7791 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.813191] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5dc4f43c-cc66-4286-8e4e-02479b7ca5e0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1233.827907] env[62753]: DEBUG nova.compute.provider_tree [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1234.070646] env[62753]: DEBUG oslo_concurrency.lockutils [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1234.331276] env[62753]: DEBUG nova.scheduler.client.report [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1234.836800] env[62753]: DEBUG oslo_concurrency.lockutils [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.677s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1234.839546] env[62753]: DEBUG oslo_concurrency.lockutils [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.769s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1234.856624] env[62753]: INFO nova.scheduler.client.report [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Deleted allocations for instance ec0b07b8-1cb9-4911-8b03-efb466a21888 [ 1235.345340] env[62753]: INFO nova.compute.claims [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1235.364116] env[62753]: DEBUG oslo_concurrency.lockutils [None req-deb67e90-ecea-4ca7-b387-7209657bb8f3 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "ec0b07b8-1cb9-4911-8b03-efb466a21888" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.267s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1235.851888] env[62753]: INFO nova.compute.resource_tracker [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating resource usage from migration 1491faad-0fb0-4f1b-8bed-510b0da4f0c1 [ 1235.936787] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-dc1e7d26-87d0-4d90-baed-35c9a25e32a2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.944476] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77edda72-493a-4e24-928d-a906abb60b43 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.976057] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a39e82-8fa0-42e7-80f1-d52f1c40bfd4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.983300] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aeff46a0-0a4e-466e-8e2b-ade7b3901d8d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1235.996732] env[62753]: DEBUG nova.compute.provider_tree [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1236.499951] env[62753]: DEBUG nova.scheduler.client.report [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1237.004873] env[62753]: DEBUG oslo_concurrency.lockutils [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.165s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1237.005267] env[62753]: INFO nova.compute.manager [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Migrating [ 1237.520445] env[62753]: DEBUG oslo_concurrency.lockutils [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1237.520686] env[62753]: DEBUG oslo_concurrency.lockutils [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1237.520909] env[62753]: DEBUG nova.network.neutron [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1238.237629] env[62753]: DEBUG nova.network.neutron [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance_info_cache with network_info: [{"id": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "address": "fa:16:3e:88:6a:1d", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25e4c4c8-9d", "ovs_interfaceid": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1238.724091] env[62753]: DEBUG oslo_vmware.rw_handles [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52171f18-abe8-bca2-e849-c523d278cfdc/disk-0.vmdk. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1238.725135] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-213be0a2-5262-4f4c-95a3-77b1de56bc0a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.730726] env[62753]: DEBUG oslo_vmware.rw_handles [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52171f18-abe8-bca2-e849-c523d278cfdc/disk-0.vmdk is in state: ready. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1238.730896] env[62753]: ERROR oslo_vmware.rw_handles [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Aborting lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52171f18-abe8-bca2-e849-c523d278cfdc/disk-0.vmdk due to incomplete transfer. [ 1238.731132] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseAbort with opID=oslo.vmware-e67f8f85-d132-4679-abc4-77235252a9c8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.737061] env[62753]: DEBUG oslo_vmware.rw_handles [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Closed VMDK read handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52171f18-abe8-bca2-e849-c523d278cfdc/disk-0.vmdk. {{(pid=62753) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:735}} [ 1238.737263] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Uploaded image 026f5cd9-1845-4b09-9364-71568c7d3e7f to the Glance image server {{(pid=62753) upload_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:472}} [ 1238.739519] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Destroying the VM {{(pid=62753) destroy_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1368}} [ 1238.739974] env[62753]: DEBUG oslo_concurrency.lockutils [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1238.741181] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Destroy_Task with opID=oslo.vmware-e10c5c24-d34d-4a00-8a0d-6586285118c3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1238.746821] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1238.746821] env[62753]: value = "task-1332970" [ 1238.746821] env[62753]: _type = "Task" [ 1238.746821] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1238.754773] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332970, 'name': Destroy_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.255585] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332970, 'name': Destroy_Task, 'duration_secs': 0.471817} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1239.255931] env[62753]: INFO nova.virt.vmwareapi.vm_util [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Destroyed the VM [ 1239.256020] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Deleting Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:932}} [ 1239.256262] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachineSnapshot.RemoveSnapshot_Task with opID=oslo.vmware-8919ef31-7896-4b68-b168-4c0b6d64d9d3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1239.263710] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1239.263710] env[62753]: value = "task-1332971" [ 1239.263710] env[62753]: _type = "Task" [ 1239.263710] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1239.270726] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332971, 'name': RemoveSnapshot_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1239.774118] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332971, 'name': RemoveSnapshot_Task} progress is 100%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.255070] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ef24e1ee-312e-4f8e-ae11-8c8a2973329a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.272818] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance '22cea165-cfde-403f-a0e2-82a86f5baa51' progress to 0 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1240.283339] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332971, 'name': RemoveSnapshot_Task, 'duration_secs': 0.755923} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.283575] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Deleted Snapshot of the VM instance {{(pid=62753) _delete_vm_snapshot /opt/stack/nova/nova/virt/vmwareapi/vmops.py:938}} [ 1240.283840] env[62753]: DEBUG nova.compute.manager [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1240.284558] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-071955f2-2f40-489f-a74e-cc28c8e5148f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.299286] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "51790f51-c464-48cc-9891-41cae5a8f63e" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1240.299517] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "51790f51-c464-48cc-9891-41cae5a8f63e" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1240.299742] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "51790f51-c464-48cc-9891-41cae5a8f63e-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1240.299932] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "51790f51-c464-48cc-9891-41cae5a8f63e-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1240.300112] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "51790f51-c464-48cc-9891-41cae5a8f63e-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1240.301841] env[62753]: INFO nova.compute.manager [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Terminating instance [ 1240.303460] env[62753]: DEBUG nova.compute.manager [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1240.303653] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1240.304388] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d76e1514-1875-4bbe-aade-526defc55224 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.311055] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1240.311845] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-c5bdb964-0283-41ad-baf9-4029c6754c4c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.318283] env[62753]: DEBUG oslo_vmware.api [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1240.318283] env[62753]: value = "task-1332972" [ 1240.318283] env[62753]: _type = "Task" [ 1240.318283] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.326297] env[62753]: DEBUG oslo_vmware.api [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332972, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.781672] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1240.781970] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-0846eec4-f5bc-409b-9646-bd8c93f94136 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.788864] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1240.788864] env[62753]: value = "task-1332973" [ 1240.788864] env[62753]: _type = "Task" [ 1240.788864] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.797596] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332973, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.798098] env[62753]: INFO nova.compute.manager [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Shelve offloading [ 1240.800060] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1240.800304] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-96664a44-b377-4c33-9a5d-79e2baabfc21 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.805657] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1240.805657] env[62753]: value = "task-1332974" [ 1240.805657] env[62753]: _type = "Task" [ 1240.805657] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.814608] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332974, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1240.826197] env[62753]: DEBUG oslo_vmware.api [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332972, 'name': PowerOffVM_Task, 'duration_secs': 0.282621} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1240.826491] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1240.826622] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1240.826999] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e8737014-b1ac-49d4-8208-1726d64cff9b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.893121] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1240.893388] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1240.893577] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Deleting the datastore file [datastore2] 51790f51-c464-48cc-9891-41cae5a8f63e {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1240.893858] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-a6824792-bd24-4b28-bdc2-9f73c4e17d5b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1240.901460] env[62753]: DEBUG oslo_vmware.api [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1240.901460] env[62753]: value = "task-1332976" [ 1240.901460] env[62753]: _type = "Task" [ 1240.901460] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1240.909154] env[62753]: DEBUG oslo_vmware.api [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332976, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.298930] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332973, 'name': PowerOffVM_Task, 'duration_secs': 0.17123} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.299292] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1241.299440] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance '22cea165-cfde-403f-a0e2-82a86f5baa51' progress to 17 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1241.314510] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] VM already powered off {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1519}} [ 1241.314751] env[62753]: DEBUG nova.compute.manager [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1241.315503] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-742b7086-2fbd-4551-8d15-7f65191d2c91 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.321163] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1241.321356] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1241.321538] env[62753]: DEBUG nova.network.neutron [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1241.411195] env[62753]: DEBUG oslo_vmware.api [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332976, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.806055] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1241.806294] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1241.806431] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1241.806611] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1241.806804] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1241.806975] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1241.807211] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1241.807377] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1241.807551] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1241.807727] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1241.807958] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1241.813091] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-978fd85a-c002-4ae8-b3c5-fd141d8ab240 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1241.832639] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1241.832639] env[62753]: value = "task-1332977" [ 1241.832639] env[62753]: _type = "Task" [ 1241.832639] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1241.844062] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332977, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1241.912357] env[62753]: DEBUG oslo_vmware.api [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332976, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.939646} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1241.912631] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1241.912855] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1241.913055] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1241.913242] env[62753]: INFO nova.compute.manager [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Took 1.61 seconds to destroy the instance on the hypervisor. [ 1241.913493] env[62753]: DEBUG oslo.service.loopingcall [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1241.913690] env[62753]: DEBUG nova.compute.manager [-] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1241.913782] env[62753]: DEBUG nova.network.neutron [-] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1242.156195] env[62753]: DEBUG nova.network.neutron [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Updating instance_info_cache with network_info: [{"id": "b81e545c-6410-4acd-8777-1c9a2ef2b7eb", "address": "fa:16:3e:12:cc:32", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb81e545c-64", "ovs_interfaceid": "b81e545c-6410-4acd-8777-1c9a2ef2b7eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.279652] env[62753]: DEBUG nova.compute.manager [req-e39bcbde-b1b9-46ff-a327-87011f11ff2b req-c10b905f-a066-42bc-bd21-d013977fd334 service nova] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Received event network-vif-deleted-450be489-a9db-47f3-91e7-c6673a8af916 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1242.279787] env[62753]: INFO nova.compute.manager [req-e39bcbde-b1b9-46ff-a327-87011f11ff2b req-c10b905f-a066-42bc-bd21-d013977fd334 service nova] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Neutron deleted interface 450be489-a9db-47f3-91e7-c6673a8af916; detaching it from the instance and deleting it from the info cache [ 1242.279953] env[62753]: DEBUG nova.network.neutron [req-e39bcbde-b1b9-46ff-a327-87011f11ff2b req-c10b905f-a066-42bc-bd21-d013977fd334 service nova] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.341873] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332977, 'name': ReconfigVM_Task, 'duration_secs': 0.258937} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1242.342208] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance '22cea165-cfde-403f-a0e2-82a86f5baa51' progress to 33 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1242.659247] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1242.759476] env[62753]: DEBUG nova.network.neutron [-] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1242.782451] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-ae90097f-536b-4013-b9f1-26d5df4f80a7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.793015] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-82175f56-35dd-48da-8859-92eaf5f7c4a2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.817689] env[62753]: DEBUG nova.compute.manager [req-e39bcbde-b1b9-46ff-a327-87011f11ff2b req-c10b905f-a066-42bc-bd21-d013977fd334 service nova] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Detach interface failed, port_id=450be489-a9db-47f3-91e7-c6673a8af916, reason: Instance 51790f51-c464-48cc-9891-41cae5a8f63e could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1242.848590] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1242.848846] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1242.849061] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1242.849447] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1242.849726] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1242.849925] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1242.850160] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1242.850498] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1242.850579] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1242.850680] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1242.850875] env[62753]: DEBUG nova.virt.hardware [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1242.856155] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Reconfiguring VM instance instance-00000068 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1242.856449] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-4b6d9877-9c57-4552-a368-5a88f6e7c651 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.875713] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1242.875713] env[62753]: value = "task-1332978" [ 1242.875713] env[62753]: _type = "Task" [ 1242.875713] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1242.883902] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332978, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1242.977151] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1242.978071] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d9927385-86f6-404c-98ce-42e0056927dc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1242.985117] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1242.985344] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-7fb8298c-008d-4301-9531-39d2795f3fed {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.044885] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1243.045137] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1243.045315] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Deleting the datastore file [datastore1] 2e784985-7045-448e-a650-f6692bb69d94 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1243.045581] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-75eefecc-87db-4a36-a68a-80db12e82e07 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1243.052021] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1243.052021] env[62753]: value = "task-1332980" [ 1243.052021] env[62753]: _type = "Task" [ 1243.052021] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1243.059630] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332980, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.262508] env[62753]: INFO nova.compute.manager [-] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Took 1.35 seconds to deallocate network for instance. [ 1243.385699] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332978, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1243.561283] env[62753]: DEBUG oslo_vmware.api [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332980, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.140306} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1243.561534] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1243.561718] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1243.561900] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1243.586414] env[62753]: INFO nova.scheduler.client.report [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Deleted allocations for instance 2e784985-7045-448e-a650-f6692bb69d94 [ 1243.768815] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1243.769123] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1243.769359] env[62753]: DEBUG nova.objects.instance [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lazy-loading 'resources' on Instance uuid 51790f51-c464-48cc-9891-41cae5a8f63e {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1243.885785] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332978, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.091148] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1244.305069] env[62753]: DEBUG nova.compute.manager [req-402e27a2-8ad8-4352-8c07-8c0c368043b3 req-2af5aa64-da75-483a-9e73-343e2a2ffe39 service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Received event network-vif-unplugged-b81e545c-6410-4acd-8777-1c9a2ef2b7eb {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1244.305228] env[62753]: DEBUG oslo_concurrency.lockutils [req-402e27a2-8ad8-4352-8c07-8c0c368043b3 req-2af5aa64-da75-483a-9e73-343e2a2ffe39 service nova] Acquiring lock "2e784985-7045-448e-a650-f6692bb69d94-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1244.305410] env[62753]: DEBUG oslo_concurrency.lockutils [req-402e27a2-8ad8-4352-8c07-8c0c368043b3 req-2af5aa64-da75-483a-9e73-343e2a2ffe39 service nova] Lock "2e784985-7045-448e-a650-f6692bb69d94-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1244.305572] env[62753]: DEBUG oslo_concurrency.lockutils [req-402e27a2-8ad8-4352-8c07-8c0c368043b3 req-2af5aa64-da75-483a-9e73-343e2a2ffe39 service nova] Lock "2e784985-7045-448e-a650-f6692bb69d94-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1244.305743] env[62753]: DEBUG nova.compute.manager [req-402e27a2-8ad8-4352-8c07-8c0c368043b3 req-2af5aa64-da75-483a-9e73-343e2a2ffe39 service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] No waiting events found dispatching network-vif-unplugged-b81e545c-6410-4acd-8777-1c9a2ef2b7eb {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1244.305925] env[62753]: WARNING nova.compute.manager [req-402e27a2-8ad8-4352-8c07-8c0c368043b3 req-2af5aa64-da75-483a-9e73-343e2a2ffe39 service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Received unexpected event network-vif-unplugged-b81e545c-6410-4acd-8777-1c9a2ef2b7eb for instance with vm_state shelved_offloaded and task_state None. [ 1244.306080] env[62753]: DEBUG nova.compute.manager [req-402e27a2-8ad8-4352-8c07-8c0c368043b3 req-2af5aa64-da75-483a-9e73-343e2a2ffe39 service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Received event network-changed-b81e545c-6410-4acd-8777-1c9a2ef2b7eb {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1244.306243] env[62753]: DEBUG nova.compute.manager [req-402e27a2-8ad8-4352-8c07-8c0c368043b3 req-2af5aa64-da75-483a-9e73-343e2a2ffe39 service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Refreshing instance network info cache due to event network-changed-b81e545c-6410-4acd-8777-1c9a2ef2b7eb. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1244.306429] env[62753]: DEBUG oslo_concurrency.lockutils [req-402e27a2-8ad8-4352-8c07-8c0c368043b3 req-2af5aa64-da75-483a-9e73-343e2a2ffe39 service nova] Acquiring lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1244.306567] env[62753]: DEBUG oslo_concurrency.lockutils [req-402e27a2-8ad8-4352-8c07-8c0c368043b3 req-2af5aa64-da75-483a-9e73-343e2a2ffe39 service nova] Acquired lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1244.306724] env[62753]: DEBUG nova.network.neutron [req-402e27a2-8ad8-4352-8c07-8c0c368043b3 req-2af5aa64-da75-483a-9e73-343e2a2ffe39 service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Refreshing network info cache for port b81e545c-6410-4acd-8777-1c9a2ef2b7eb {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1244.338941] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f1f2d221-d9c0-4df2-8250-1a22382ed3fa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.348881] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c3fba62d-4927-4f36-a832-5629ea58f650 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.378236] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4aaaffc1-bf9e-4c73-a663-c044b3397256 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.390634] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-077f3034-c43a-4850-a788-2c38ba081b3f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.394520] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332978, 'name': ReconfigVM_Task, 'duration_secs': 1.174352} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.394854] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Reconfigured VM instance instance-00000068 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1244.395953] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bd7bcfb3-91e5-4ffd-94cd-1ce176d12b84 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.406274] env[62753]: DEBUG nova.compute.provider_tree [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1244.427229] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Reconfiguring VM instance instance-00000068 to attach disk [datastore1] 22cea165-cfde-403f-a0e2-82a86f5baa51/22cea165-cfde-403f-a0e2-82a86f5baa51.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1244.428106] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-f6b03afd-ae7d-4a43-b38c-be6f7349087a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1244.446419] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1244.446419] env[62753]: value = "task-1332981" [ 1244.446419] env[62753]: _type = "Task" [ 1244.446419] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1244.454495] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332981, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1244.909482] env[62753]: DEBUG nova.scheduler.client.report [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1244.958224] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332981, 'name': ReconfigVM_Task, 'duration_secs': 0.295311} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1244.958390] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Reconfigured VM instance instance-00000068 to attach disk [datastore1] 22cea165-cfde-403f-a0e2-82a86f5baa51/22cea165-cfde-403f-a0e2-82a86f5baa51.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1244.958839] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance '22cea165-cfde-403f-a0e2-82a86f5baa51' progress to 50 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1245.025783] env[62753]: DEBUG nova.network.neutron [req-402e27a2-8ad8-4352-8c07-8c0c368043b3 req-2af5aa64-da75-483a-9e73-343e2a2ffe39 service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Updated VIF entry in instance network info cache for port b81e545c-6410-4acd-8777-1c9a2ef2b7eb. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1245.026123] env[62753]: DEBUG nova.network.neutron [req-402e27a2-8ad8-4352-8c07-8c0c368043b3 req-2af5aa64-da75-483a-9e73-343e2a2ffe39 service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Updating instance_info_cache with network_info: [{"id": "b81e545c-6410-4acd-8777-1c9a2ef2b7eb", "address": "fa:16:3e:12:cc:32", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": null, "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "unbound", "details": {}, "devname": "tapb81e545c-64", "ovs_interfaceid": null, "qbh_params": null, "qbg_params": null, "active": false, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1245.416106] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.647s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1245.418483] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 1.327s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1245.418719] env[62753]: DEBUG nova.objects.instance [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lazy-loading 'resources' on Instance uuid 2e784985-7045-448e-a650-f6692bb69d94 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1245.436258] env[62753]: INFO nova.scheduler.client.report [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Deleted allocations for instance 51790f51-c464-48cc-9891-41cae5a8f63e [ 1245.465384] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb8233a8-0b1f-4c00-8245-5817134208ab {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.485071] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-168f8128-73b1-4cd7-90f8-e3d8217a0306 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1245.502429] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance '22cea165-cfde-403f-a0e2-82a86f5baa51' progress to 67 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1245.522376] env[62753]: DEBUG oslo_concurrency.lockutils [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "2e784985-7045-448e-a650-f6692bb69d94" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1245.528234] env[62753]: DEBUG oslo_concurrency.lockutils [req-402e27a2-8ad8-4352-8c07-8c0c368043b3 req-2af5aa64-da75-483a-9e73-343e2a2ffe39 service nova] Releasing lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1245.921381] env[62753]: DEBUG nova.objects.instance [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lazy-loading 'numa_topology' on Instance uuid 2e784985-7045-448e-a650-f6692bb69d94 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1245.943827] env[62753]: DEBUG oslo_concurrency.lockutils [None req-bd0bbfcf-929c-49ca-a736-eceb9b206eb6 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "51790f51-c464-48cc-9891-41cae5a8f63e" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.644s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1246.047338] env[62753]: DEBUG nova.network.neutron [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Port 25e4c4c8-9d78-4a88-9a5e-4a364ee574a5 binding to destination host cpu-1 is already ACTIVE {{(pid=62753) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1246.423658] env[62753]: DEBUG nova.objects.base [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Object Instance<2e784985-7045-448e-a650-f6692bb69d94> lazy-loaded attributes: resources,numa_topology {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1246.490888] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f87a66f4-46f7-4f31-bb2d-77fb8b812602 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.498581] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9e2cac30-05f8-4c74-81dd-d50fdebfcfc3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.527690] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5772e4f9-44bc-4b3d-86f7-29447f2a206c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.534362] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-31c78f55-d139-4509-a2f7-738ee9772299 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.546971] env[62753]: DEBUG nova.compute.provider_tree [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1246.880200] env[62753]: DEBUG oslo_concurrency.lockutils [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "a2e7fa98-566e-40f1-8da7-9318de89c14c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1246.880449] env[62753]: DEBUG oslo_concurrency.lockutils [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "a2e7fa98-566e-40f1-8da7-9318de89c14c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1246.880666] env[62753]: DEBUG oslo_concurrency.lockutils [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "a2e7fa98-566e-40f1-8da7-9318de89c14c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1246.880854] env[62753]: DEBUG oslo_concurrency.lockutils [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "a2e7fa98-566e-40f1-8da7-9318de89c14c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1246.881040] env[62753]: DEBUG oslo_concurrency.lockutils [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "a2e7fa98-566e-40f1-8da7-9318de89c14c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1246.883253] env[62753]: INFO nova.compute.manager [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Terminating instance [ 1246.884930] env[62753]: DEBUG nova.compute.manager [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1246.885150] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1246.885984] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-94cbdb40-f5ce-42be-b119-de2f6b1ff59d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.893833] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1246.894064] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-d632fe96-287d-43ef-bead-4499a65e6015 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1246.900186] env[62753]: DEBUG oslo_vmware.api [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1246.900186] env[62753]: value = "task-1332982" [ 1246.900186] env[62753]: _type = "Task" [ 1246.900186] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1246.908737] env[62753]: DEBUG oslo_vmware.api [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332982, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.049774] env[62753]: DEBUG nova.scheduler.client.report [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1247.066910] env[62753]: DEBUG oslo_concurrency.lockutils [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "22cea165-cfde-403f-a0e2-82a86f5baa51-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1247.067144] env[62753]: DEBUG oslo_concurrency.lockutils [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1247.067319] env[62753]: DEBUG oslo_concurrency.lockutils [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1247.409677] env[62753]: DEBUG oslo_vmware.api [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332982, 'name': PowerOffVM_Task, 'duration_secs': 0.177946} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.409936] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1247.410123] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1247.410369] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-37e8be63-9cb9-45d4-ba45-eb5101c17f11 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.470200] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1247.470586] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1247.470637] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Deleting the datastore file [datastore2] a2e7fa98-566e-40f1-8da7-9318de89c14c {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1247.470857] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-763864d3-7fb5-4b03-9526-dc0fcb37e1e6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1247.477478] env[62753]: DEBUG oslo_vmware.api [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for the task: (returnval){ [ 1247.477478] env[62753]: value = "task-1332984" [ 1247.477478] env[62753]: _type = "Task" [ 1247.477478] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1247.484926] env[62753]: DEBUG oslo_vmware.api [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332984, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1247.555041] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 2.136s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1247.987885] env[62753]: DEBUG oslo_vmware.api [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Task: {'id': task-1332984, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146331} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1247.988162] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1247.988465] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1247.988657] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1247.988837] env[62753]: INFO nova.compute.manager [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Took 1.10 seconds to destroy the instance on the hypervisor. [ 1247.989108] env[62753]: DEBUG oslo.service.loopingcall [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1247.989330] env[62753]: DEBUG nova.compute.manager [-] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1247.989420] env[62753]: DEBUG nova.network.neutron [-] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1248.009639] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1248.009840] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1248.009978] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62753) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1248.062202] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9a0a3521-9e7e-4aac-9ab0-77b9021c1166 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "2e784985-7045-448e-a650-f6692bb69d94" "released" by "nova.compute.manager.ComputeManager.shelve_instance..do_shelve_instance" :: held 23.610s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1248.063089] env[62753]: DEBUG oslo_concurrency.lockutils [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "2e784985-7045-448e-a650-f6692bb69d94" acquired by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: waited 2.541s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1248.063264] env[62753]: INFO nova.compute.manager [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Unshelving [ 1248.116854] env[62753]: DEBUG oslo_concurrency.lockutils [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1248.117122] env[62753]: DEBUG oslo_concurrency.lockutils [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1248.117309] env[62753]: DEBUG nova.network.neutron [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1248.271995] env[62753]: DEBUG nova.compute.manager [req-26155d20-3265-4fc7-b317-d85d968aa7ef req-6f6beeb6-e291-4e3b-aa79-630a5a6da550 service nova] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Received event network-vif-deleted-dd52d78f-614d-4686-a0aa-f92598ea6932 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1248.272207] env[62753]: INFO nova.compute.manager [req-26155d20-3265-4fc7-b317-d85d968aa7ef req-6f6beeb6-e291-4e3b-aa79-630a5a6da550 service nova] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Neutron deleted interface dd52d78f-614d-4686-a0aa-f92598ea6932; detaching it from the instance and deleting it from the info cache [ 1248.272383] env[62753]: DEBUG nova.network.neutron [req-26155d20-3265-4fc7-b317-d85d968aa7ef req-6f6beeb6-e291-4e3b-aa79-630a5a6da550 service nova] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1248.747118] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "221f76f0-ee69-4119-b19b-c148e40c1126" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1248.747412] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "221f76f0-ee69-4119-b19b-c148e40c1126" acquired by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1248.748576] env[62753]: DEBUG nova.network.neutron [-] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1248.774171] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-4e59573e-50d2-4b25-954a-46602bf65e8e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.783236] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fe7ac231-9e5f-45b2-8472-9d987cdfaaf8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1248.806259] env[62753]: DEBUG nova.compute.manager [req-26155d20-3265-4fc7-b317-d85d968aa7ef req-6f6beeb6-e291-4e3b-aa79-630a5a6da550 service nova] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Detach interface failed, port_id=dd52d78f-614d-4686-a0aa-f92598ea6932, reason: Instance a2e7fa98-566e-40f1-8da7-9318de89c14c could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1248.828439] env[62753]: DEBUG nova.network.neutron [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance_info_cache with network_info: [{"id": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "address": "fa:16:3e:88:6a:1d", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25e4c4c8-9d", "ovs_interfaceid": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1249.010160] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1249.010388] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1249.089397] env[62753]: DEBUG oslo_concurrency.lockutils [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.089397] env[62753]: DEBUG oslo_concurrency.lockutils [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1249.089601] env[62753]: DEBUG nova.objects.instance [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lazy-loading 'pci_requests' on Instance uuid 2e784985-7045-448e-a650-f6692bb69d94 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1249.252108] env[62753]: INFO nova.compute.manager [-] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Took 1.26 seconds to deallocate network for instance. [ 1249.253060] env[62753]: DEBUG nova.compute.utils [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1249.330655] env[62753]: DEBUG oslo_concurrency.lockutils [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1249.593685] env[62753]: DEBUG nova.objects.instance [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lazy-loading 'numa_topology' on Instance uuid 2e784985-7045-448e-a650-f6692bb69d94 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1249.757661] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "221f76f0-ee69-4119-b19b-c148e40c1126" "released" by "nova.compute.manager.ComputeManager.reserve_block_device_name..do_reserve" :: held 1.010s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1249.759491] env[62753]: DEBUG oslo_concurrency.lockutils [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1249.851182] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-bc40073e-f486-4b24-8c86-2c52bce9f106 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.870711] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-72bab229-7dc5-4e1d-84d3-3f1af0b29492 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1249.877276] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance '22cea165-cfde-403f-a0e2-82a86f5baa51' progress to 83 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1250.010609] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1250.095849] env[62753]: INFO nova.compute.claims [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1250.383496] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1250.383967] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-d943096b-ecb9-434b-b23e-1dcf9ab9806d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.391452] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1250.391452] env[62753]: value = "task-1332985" [ 1250.391452] env[62753]: _type = "Task" [ 1250.391452] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1250.399257] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332985, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1250.810754] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "221f76f0-ee69-4119-b19b-c148e40c1126" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1250.811067] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "221f76f0-ee69-4119-b19b-c148e40c1126" acquired by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1250.811314] env[62753]: INFO nova.compute.manager [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Attaching volume 95aec84e-bc3f-4074-b379-2b5bf5976d96 to /dev/sdb [ 1250.840386] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6aa3b4e7-c7ed-4ad0-a758-012bb1f8226b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.847501] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fa849158-e69e-4e75-9256-064b6958a4a2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1250.860401] env[62753]: DEBUG nova.virt.block_device [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Updating existing volume attachment record: 208ab119-9b09-429f-a4f9-bc4b728c9299 {{(pid=62753) _volume_attach /opt/stack/nova/nova/virt/block_device.py:666}} [ 1250.900890] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332985, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1251.010627] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1251.169152] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8827d789-9bde-42ec-b1ad-2e39a5c3a4ab {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.176800] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b3ac2ba0-0db6-478f-b552-cc5dd39e02ed {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.208024] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-03777f95-3b79-4170-bae6-ea70c992ec3c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.215211] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b1d41afb-5934-4be3-82d7-49058867d645 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1251.871012] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1251.884121] env[62753]: DEBUG nova.compute.provider_tree [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1251.888124] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332985, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.380091] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332985, 'name': PowerOnVM_Task} progress is 66%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1252.390168] env[62753]: DEBUG nova.scheduler.client.report [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1252.881408] env[62753]: DEBUG oslo_vmware.api [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332985, 'name': PowerOnVM_Task, 'duration_secs': 2.095998} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1252.881690] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1252.881892] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-be626c83-22e1-4fa5-bef0-c71d197c8744 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance '22cea165-cfde-403f-a0e2-82a86f5baa51' progress to 100 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1252.894844] env[62753]: DEBUG oslo_concurrency.lockutils [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 3.805s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1252.897112] env[62753]: DEBUG oslo_concurrency.lockutils [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 3.138s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1252.897340] env[62753]: DEBUG nova.objects.instance [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lazy-loading 'resources' on Instance uuid a2e7fa98-566e-40f1-8da7-9318de89c14c {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1252.928580] env[62753]: INFO nova.network.neutron [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Updating port b81e545c-6410-4acd-8777-1c9a2ef2b7eb with attributes {'binding:host_id': 'cpu-1', 'device_owner': 'compute:nova'} [ 1253.466732] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ac5f3c4-7f03-4c92-bc0a-fb1f7029eade {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.474744] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3f5a387f-426c-45e5-84b4-5593894ea9a0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.504219] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8df5e7ea-83d4-404f-b705-02f7f2359f1e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.510905] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6ba0ec60-126e-4cb2-a6ca-c796dff41df6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1253.523234] env[62753]: DEBUG nova.compute.provider_tree [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1254.026660] env[62753]: DEBUG nova.scheduler.client.report [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1254.323974] env[62753]: DEBUG nova.compute.manager [req-7316747e-b6bf-40de-b988-fbe20ed581c6 req-e8ad2e56-dab4-47f4-99d4-b5ae3c4eab91 service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Received event network-vif-plugged-b81e545c-6410-4acd-8777-1c9a2ef2b7eb {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1254.324235] env[62753]: DEBUG oslo_concurrency.lockutils [req-7316747e-b6bf-40de-b988-fbe20ed581c6 req-e8ad2e56-dab4-47f4-99d4-b5ae3c4eab91 service nova] Acquiring lock "2e784985-7045-448e-a650-f6692bb69d94-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1254.324440] env[62753]: DEBUG oslo_concurrency.lockutils [req-7316747e-b6bf-40de-b988-fbe20ed581c6 req-e8ad2e56-dab4-47f4-99d4-b5ae3c4eab91 service nova] Lock "2e784985-7045-448e-a650-f6692bb69d94-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1254.324633] env[62753]: DEBUG oslo_concurrency.lockutils [req-7316747e-b6bf-40de-b988-fbe20ed581c6 req-e8ad2e56-dab4-47f4-99d4-b5ae3c4eab91 service nova] Lock "2e784985-7045-448e-a650-f6692bb69d94-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1254.324873] env[62753]: DEBUG nova.compute.manager [req-7316747e-b6bf-40de-b988-fbe20ed581c6 req-e8ad2e56-dab4-47f4-99d4-b5ae3c4eab91 service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] No waiting events found dispatching network-vif-plugged-b81e545c-6410-4acd-8777-1c9a2ef2b7eb {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1254.325056] env[62753]: WARNING nova.compute.manager [req-7316747e-b6bf-40de-b988-fbe20ed581c6 req-e8ad2e56-dab4-47f4-99d4-b5ae3c4eab91 service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Received unexpected event network-vif-plugged-b81e545c-6410-4acd-8777-1c9a2ef2b7eb for instance with vm_state shelved_offloaded and task_state spawning. [ 1254.422535] env[62753]: DEBUG oslo_concurrency.lockutils [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1254.422535] env[62753]: DEBUG oslo_concurrency.lockutils [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1254.422535] env[62753]: DEBUG nova.network.neutron [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1254.532564] env[62753]: DEBUG oslo_concurrency.lockutils [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.635s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1254.534758] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 2.664s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1254.534958] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1254.535136] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62753) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1254.536012] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-97da54f8-f6a3-4e69-ae63-5a4dffe07801 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.544325] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3c71e68c-3ce0-4b2d-b054-120b9e7a46e6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.559145] env[62753]: INFO nova.scheduler.client.report [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Deleted allocations for instance a2e7fa98-566e-40f1-8da7-9318de89c14c [ 1254.560983] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6dda15c0-124a-41e6-9d1b-5b7715f45939 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.569710] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2bcc218c-0fc2-45d4-90cf-7bdcc7cf4c6c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1254.601179] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=180384MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=62753) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1254.601329] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1254.601514] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1255.071077] env[62753]: DEBUG oslo_concurrency.lockutils [None req-adc1c8a7-0f48-4398-aba0-f73576a00083 tempest-ServerRescueNegativeTestJSON-1867183818 tempest-ServerRescueNegativeTestJSON-1867183818-project-member] Lock "a2e7fa98-566e-40f1-8da7-9318de89c14c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 8.191s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1255.109380] env[62753]: DEBUG oslo_concurrency.lockutils [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "22cea165-cfde-403f-a0e2-82a86f5baa51" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1255.109726] env[62753]: DEBUG oslo_concurrency.lockutils [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51" acquired by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1255.109919] env[62753]: DEBUG nova.compute.manager [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Going to confirm migration 6 {{(pid=62753) do_confirm_resize /opt/stack/nova/nova/compute/manager.py:4783}} [ 1255.135677] env[62753]: DEBUG nova.network.neutron [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Updating instance_info_cache with network_info: [{"id": "b81e545c-6410-4acd-8777-1c9a2ef2b7eb", "address": "fa:16:3e:12:cc:32", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb81e545c-64", "ovs_interfaceid": "b81e545c-6410-4acd-8777-1c9a2ef2b7eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1255.406498] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Volume attach. Driver type: vmdk {{(pid=62753) attach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:439}} [ 1255.406749] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] _attach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284708', 'volume_id': '95aec84e-bc3f-4074-b379-2b5bf5976d96', 'name': 'volume-95aec84e-bc3f-4074-b379-2b5bf5976d96', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '221f76f0-ee69-4119-b19b-c148e40c1126', 'attached_at': '', 'detached_at': '', 'volume_id': '95aec84e-bc3f-4074-b379-2b5bf5976d96', 'serial': '95aec84e-bc3f-4074-b379-2b5bf5976d96'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:336}} [ 1255.408006] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-21c78793-6a5f-4eaf-bdb7-eba66def7c1b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.425517] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-77b41ae8-2950-4615-b41e-e2743fe26949 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.450870] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Reconfiguring VM instance instance-0000006f to attach disk [datastore2] volume-95aec84e-bc3f-4074-b379-2b5bf5976d96/volume-95aec84e-bc3f-4074-b379-2b5bf5976d96.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1255.451208] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d461de2f-9691-4e1b-8c3b-4cfdee986446 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.469785] env[62753]: DEBUG oslo_vmware.api [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1255.469785] env[62753]: value = "task-1332988" [ 1255.469785] env[62753]: _type = "Task" [ 1255.469785] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.477793] env[62753]: DEBUG oslo_vmware.api [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332988, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.612403] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Applying migration context for instance 22cea165-cfde-403f-a0e2-82a86f5baa51 as it has an incoming, in-progress migration 1491faad-0fb0-4f1b-8bed-510b0da4f0c1. Migration status is confirming {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1016}} [ 1255.613282] env[62753]: INFO nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating resource usage from migration 1491faad-0fb0-4f1b-8bed-510b0da4f0c1 [ 1255.634458] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 221f76f0-ee69-4119-b19b-c148e40c1126 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1255.634658] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Migration 1491faad-0fb0-4f1b-8bed-510b0da4f0c1 is active on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1712}} [ 1255.634807] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 22cea165-cfde-403f-a0e2-82a86f5baa51 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1255.634962] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 2e784985-7045-448e-a650-f6692bb69d94 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1255.635211] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Total usable vcpus: 48, total allocated vcpus: 4 {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1255.635392] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=1344MB phys_disk=200GB used_disk=4GB total_vcpus=48 used_vcpus=4 pci_stats=[] {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1255.638621] env[62753]: DEBUG oslo_concurrency.lockutils [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1255.672859] env[62753]: DEBUG nova.virt.hardware [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='70743154cf72e2a74e86e7bdc23c5be5',container_format='bare',created_at=2024-10-21T20:37:28Z,direct_url=,disk_format='vmdk',id=026f5cd9-1845-4b09-9364-71568c7d3e7f,min_disk=1,min_ram=0,name='tempest-ServerActionsTestOtherB-server-692598728-shelved',owner='196457eaa51643c38b4ef4fc6fbbcba3',properties=ImageMetaProps,protected=,size=31669248,status='active',tags=,updated_at=2024-10-21T20:37:43Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1255.673143] env[62753]: DEBUG nova.virt.hardware [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1255.673327] env[62753]: DEBUG nova.virt.hardware [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1255.673508] env[62753]: DEBUG nova.virt.hardware [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1255.673659] env[62753]: DEBUG nova.virt.hardware [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1255.673810] env[62753]: DEBUG nova.virt.hardware [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1255.674056] env[62753]: DEBUG nova.virt.hardware [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1255.674243] env[62753]: DEBUG nova.virt.hardware [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1255.674425] env[62753]: DEBUG nova.virt.hardware [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1255.674593] env[62753]: DEBUG nova.virt.hardware [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1255.674769] env[62753]: DEBUG nova.virt.hardware [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1255.675759] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-727267e8-ef36-4730-bd90-0bb42bd8340a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.686178] env[62753]: DEBUG oslo_concurrency.lockutils [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1255.686178] env[62753]: DEBUG oslo_concurrency.lockutils [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1255.686178] env[62753]: DEBUG nova.network.neutron [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1255.686178] env[62753]: DEBUG nova.objects.instance [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lazy-loading 'info_cache' on Instance uuid 22cea165-cfde-403f-a0e2-82a86f5baa51 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1255.691566] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55d494b1-8349-40c0-9eba-9d976ecd569a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.708355] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:12:cc:32', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'a92a4ffe-7939-4697-bf98-5b22e2c7feda', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': 'b81e545c-6410-4acd-8777-1c9a2ef2b7eb', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1255.716996] env[62753]: DEBUG oslo.service.loopingcall [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1255.717645] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1255.717887] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-4dffe581-2c3b-44b1-bf69-2af13e98ee17 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.742133] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1255.742133] env[62753]: value = "task-1332989" [ 1255.742133] env[62753]: _type = "Task" [ 1255.742133] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1255.750911] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332989, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1255.752775] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8096189f-7fcb-46e0-b4ea-fd17d27bbae8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.759988] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0be9d7b8-d189-429b-9b9d-0285da1481f5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.792196] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-566f6cf3-407f-467d-97c6-20e571b021af {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.799891] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-36378f04-ef65-4263-99bb-f735facbe593 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1255.813176] env[62753]: DEBUG nova.compute.provider_tree [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1255.980955] env[62753]: DEBUG oslo_vmware.api [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332988, 'name': ReconfigVM_Task, 'duration_secs': 0.386839} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1255.981267] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Reconfigured VM instance instance-0000006f to attach disk [datastore2] volume-95aec84e-bc3f-4074-b379-2b5bf5976d96/volume-95aec84e-bc3f-4074-b379-2b5bf5976d96.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1255.985950] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-b816f66f-cd98-44e9-8765-88871de88e7a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.001165] env[62753]: DEBUG oslo_vmware.api [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1256.001165] env[62753]: value = "task-1332990" [ 1256.001165] env[62753]: _type = "Task" [ 1256.001165] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.009154] env[62753]: DEBUG oslo_vmware.api [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332990, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.251710] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1332989, 'name': CreateVM_Task, 'duration_secs': 0.340423} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.251988] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1256.252669] env[62753]: DEBUG oslo_concurrency.lockutils [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/026f5cd9-1845-4b09-9364-71568c7d3e7f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1256.252750] env[62753]: DEBUG oslo_concurrency.lockutils [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "[datastore2] devstack-image-cache_base/026f5cd9-1845-4b09-9364-71568c7d3e7f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.253136] env[62753]: DEBUG oslo_concurrency.lockutils [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/026f5cd9-1845-4b09-9364-71568c7d3e7f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1256.253400] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c7c582b1-9850-4474-add6-db2799e72190 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.257799] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1256.257799] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e6b354-d4a6-9fab-5483-0aca48063404" [ 1256.257799] env[62753]: _type = "Task" [ 1256.257799] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.265727] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e6b354-d4a6-9fab-5483-0aca48063404, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.316158] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1256.352207] env[62753]: DEBUG nova.compute.manager [req-b83b1561-443c-4233-9484-ce91a6f9b7db req-c9f38258-046e-4adc-bacc-decf4dbc233d service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Received event network-changed-b81e545c-6410-4acd-8777-1c9a2ef2b7eb {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1256.352483] env[62753]: DEBUG nova.compute.manager [req-b83b1561-443c-4233-9484-ce91a6f9b7db req-c9f38258-046e-4adc-bacc-decf4dbc233d service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Refreshing instance network info cache due to event network-changed-b81e545c-6410-4acd-8777-1c9a2ef2b7eb. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1256.352610] env[62753]: DEBUG oslo_concurrency.lockutils [req-b83b1561-443c-4233-9484-ce91a6f9b7db req-c9f38258-046e-4adc-bacc-decf4dbc233d service nova] Acquiring lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1256.352735] env[62753]: DEBUG oslo_concurrency.lockutils [req-b83b1561-443c-4233-9484-ce91a6f9b7db req-c9f38258-046e-4adc-bacc-decf4dbc233d service nova] Acquired lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.352901] env[62753]: DEBUG nova.network.neutron [req-b83b1561-443c-4233-9484-ce91a6f9b7db req-c9f38258-046e-4adc-bacc-decf4dbc233d service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Refreshing network info cache for port b81e545c-6410-4acd-8777-1c9a2ef2b7eb {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1256.510760] env[62753]: DEBUG oslo_vmware.api [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332990, 'name': ReconfigVM_Task, 'duration_secs': 0.135602} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1256.510760] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Attached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284708', 'volume_id': '95aec84e-bc3f-4074-b379-2b5bf5976d96', 'name': 'volume-95aec84e-bc3f-4074-b379-2b5bf5976d96', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '221f76f0-ee69-4119-b19b-c148e40c1126', 'attached_at': '', 'detached_at': '', 'volume_id': '95aec84e-bc3f-4074-b379-2b5bf5976d96', 'serial': '95aec84e-bc3f-4074-b379-2b5bf5976d96'} {{(pid=62753) _attach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:361}} [ 1256.768054] env[62753]: DEBUG oslo_concurrency.lockutils [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "[datastore2] devstack-image-cache_base/026f5cd9-1845-4b09-9364-71568c7d3e7f" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1256.771171] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Processing image 026f5cd9-1845-4b09-9364-71568c7d3e7f {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1256.771171] env[62753]: DEBUG oslo_concurrency.lockutils [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/026f5cd9-1845-4b09-9364-71568c7d3e7f/026f5cd9-1845-4b09-9364-71568c7d3e7f.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1256.771171] env[62753]: DEBUG oslo_concurrency.lockutils [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquired lock "[datastore2] devstack-image-cache_base/026f5cd9-1845-4b09-9364-71568c7d3e7f/026f5cd9-1845-4b09-9364-71568c7d3e7f.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1256.771171] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1256.771171] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-0941c88f-020d-4afb-a418-47925e13a7f8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.786844] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1256.787081] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1256.787833] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-8c8a1b97-b5d5-47be-acbf-35f446b02764 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1256.792807] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1256.792807] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5233fd7f-6556-2ce6-eea5-e011de492da1" [ 1256.792807] env[62753]: _type = "Task" [ 1256.792807] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1256.800813] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]5233fd7f-6556-2ce6-eea5-e011de492da1, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1256.821681] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1256.821869] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.220s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1256.989128] env[62753]: DEBUG nova.network.neutron [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance_info_cache with network_info: [{"id": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "address": "fa:16:3e:88:6a:1d", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.3", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap25e4c4c8-9d", "ovs_interfaceid": "25e4c4c8-9d78-4a88-9a5e-4a364ee574a5", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1257.122514] env[62753]: DEBUG nova.network.neutron [req-b83b1561-443c-4233-9484-ce91a6f9b7db req-c9f38258-046e-4adc-bacc-decf4dbc233d service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Updated VIF entry in instance network info cache for port b81e545c-6410-4acd-8777-1c9a2ef2b7eb. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1257.123053] env[62753]: DEBUG nova.network.neutron [req-b83b1561-443c-4233-9484-ce91a6f9b7db req-c9f38258-046e-4adc-bacc-decf4dbc233d service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Updating instance_info_cache with network_info: [{"id": "b81e545c-6410-4acd-8777-1c9a2ef2b7eb", "address": "fa:16:3e:12:cc:32", "network": {"id": "7341c5a5-d508-4ad8-8a67-0a90416d9bdd", "bridge": "br-int", "label": "tempest-ServerActionsTestOtherB-1851670500-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.13", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.203", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "196457eaa51643c38b4ef4fc6fbbcba3", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "a92a4ffe-7939-4697-bf98-5b22e2c7feda", "external-id": "nsx-vlan-transportzone-732", "segmentation_id": 732, "bound_drivers": {"0": "nsxv3"}}, "devname": "tapb81e545c-64", "ovs_interfaceid": "b81e545c-6410-4acd-8777-1c9a2ef2b7eb", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1257.303985] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Preparing fetch location {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:633}} [ 1257.304372] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Fetch image to [datastore2] OSTACK_IMG_f876bfa5-fcb8-4250-a141-f794a40f5863/OSTACK_IMG_f876bfa5-fcb8-4250-a141-f794a40f5863.vmdk {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:635}} [ 1257.304416] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Downloading stream optimized image 026f5cd9-1845-4b09-9364-71568c7d3e7f to [datastore2] OSTACK_IMG_f876bfa5-fcb8-4250-a141-f794a40f5863/OSTACK_IMG_f876bfa5-fcb8-4250-a141-f794a40f5863.vmdk on the data store datastore2 as vApp {{(pid=62753) _fetch_image_as_vapp /opt/stack/nova/nova/virt/vmwareapi/vmops.py:437}} [ 1257.304564] env[62753]: DEBUG nova.virt.vmwareapi.images [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Downloading image file data 026f5cd9-1845-4b09-9364-71568c7d3e7f to the ESX as VM named 'OSTACK_IMG_f876bfa5-fcb8-4250-a141-f794a40f5863' {{(pid=62753) fetch_image_stream_optimized /opt/stack/nova/nova/virt/vmwareapi/images.py:336}} [ 1257.374449] env[62753]: DEBUG oslo_vmware.rw_handles [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Creating HttpNfcLease lease for vApp import into resource pool: (val){ [ 1257.374449] env[62753]: value = "resgroup-9" [ 1257.374449] env[62753]: _type = "ResourcePool" [ 1257.374449] env[62753]: }. {{(pid=62753) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:453}} [ 1257.374772] env[62753]: DEBUG oslo_vmware.service [-] Invoking ResourcePool.ImportVApp with opID=oslo.vmware-18db05ff-9954-4699-a5e3-5272cfe78550 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.395877] env[62753]: DEBUG oslo_vmware.rw_handles [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lease: (returnval){ [ 1257.395877] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e5948d-78d7-a750-38fe-34e6dcea5a6b" [ 1257.395877] env[62753]: _type = "HttpNfcLease" [ 1257.395877] env[62753]: } obtained for vApp import into resource pool (val){ [ 1257.395877] env[62753]: value = "resgroup-9" [ 1257.395877] env[62753]: _type = "ResourcePool" [ 1257.395877] env[62753]: }. {{(pid=62753) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:461}} [ 1257.396313] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the lease: (returnval){ [ 1257.396313] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e5948d-78d7-a750-38fe-34e6dcea5a6b" [ 1257.396313] env[62753]: _type = "HttpNfcLease" [ 1257.396313] env[62753]: } to be ready. {{(pid=62753) wait_for_lease_ready /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:462}} [ 1257.401872] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1257.401872] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e5948d-78d7-a750-38fe-34e6dcea5a6b" [ 1257.401872] env[62753]: _type = "HttpNfcLease" [ 1257.401872] env[62753]: } is initializing. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:490}} [ 1257.491784] env[62753]: DEBUG oslo_concurrency.lockutils [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "refresh_cache-22cea165-cfde-403f-a0e2-82a86f5baa51" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1257.492172] env[62753]: DEBUG nova.objects.instance [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lazy-loading 'migration_context' on Instance uuid 22cea165-cfde-403f-a0e2-82a86f5baa51 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1257.550047] env[62753]: DEBUG nova.objects.instance [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lazy-loading 'flavor' on Instance uuid 221f76f0-ee69-4119-b19b-c148e40c1126 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1257.626535] env[62753]: DEBUG oslo_concurrency.lockutils [req-b83b1561-443c-4233-9484-ce91a6f9b7db req-c9f38258-046e-4adc-bacc-decf4dbc233d service nova] Releasing lock "refresh_cache-2e784985-7045-448e-a650-f6692bb69d94" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1257.817323] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1257.817501] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1257.903871] env[62753]: DEBUG oslo_vmware.api [-] Lease: (returnval){ [ 1257.903871] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e5948d-78d7-a750-38fe-34e6dcea5a6b" [ 1257.903871] env[62753]: _type = "HttpNfcLease" [ 1257.903871] env[62753]: } is ready. {{(pid=62753) _poll_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:487}} [ 1257.904241] env[62753]: DEBUG oslo_vmware.rw_handles [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Invoking VIM API for reading info of lease: (returnval){ [ 1257.904241] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e5948d-78d7-a750-38fe-34e6dcea5a6b" [ 1257.904241] env[62753]: _type = "HttpNfcLease" [ 1257.904241] env[62753]: }. {{(pid=62753) _create_import_vapp_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:467}} [ 1257.905007] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d615af3a-d346-4234-89b7-43122cb0ae2b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.912188] env[62753]: DEBUG oslo_vmware.rw_handles [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Found VMDK URL: https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52128711-3166-1e28-e12c-541efbcac951/disk-0.vmdk from lease info. {{(pid=62753) _find_vmdk_url /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:531}} [ 1257.912387] env[62753]: DEBUG oslo_vmware.rw_handles [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Creating HTTP connection to write to file with size = 31669248 and URL = https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52128711-3166-1e28-e12c-541efbcac951/disk-0.vmdk. {{(pid=62753) _create_write_connection /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:122}} [ 1257.977617] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseProgress with opID=oslo.vmware-39c666d2-a6f9-426c-8d71-f44ea5774c1b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1257.995545] env[62753]: DEBUG nova.objects.base [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Object Instance<22cea165-cfde-403f-a0e2-82a86f5baa51> lazy-loaded attributes: info_cache,migration_context {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1257.996468] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ba6948e3-8a9e-4d3e-bd3c-792622e6d401 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.014026] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-b605be70-8230-43de-b077-0281aedb24bd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.019095] env[62753]: DEBUG oslo_vmware.api [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1258.019095] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f7a7ce-fd03-38dd-599f-26f787f6659a" [ 1258.019095] env[62753]: _type = "Task" [ 1258.019095] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.026440] env[62753]: DEBUG oslo_vmware.api [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52f7a7ce-fd03-38dd-599f-26f787f6659a, 'name': SearchDatastore_Task} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1258.026675] env[62753]: DEBUG oslo_concurrency.lockutils [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1258.026907] env[62753]: DEBUG oslo_concurrency.lockutils [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1258.053999] env[62753]: DEBUG oslo_concurrency.lockutils [None req-c1d49af0-8033-41fc-92ee-bd92460d6aa8 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "221f76f0-ee69-4119-b19b-c148e40c1126" "released" by "nova.compute.manager.ComputeManager.attach_volume..do_attach_volume" :: held 7.243s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1258.260525] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "221f76f0-ee69-4119-b19b-c148e40c1126" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1258.260844] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "221f76f0-ee69-4119-b19b-c148e40c1126" acquired by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1258.323347] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1258.323746] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Starting heal instance info cache {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1258.601188] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3aee7ada-6dad-4327-a155-5c6ad31482f6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.612763] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-12058736-6c77-4039-92a7-825cfc4d5e6d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.648074] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a341b5c-ee37-47cf-babb-e8770c663012 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.657413] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40e1823-6753-4469-891d-cb9ba532287d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.673319] env[62753]: DEBUG nova.compute.provider_tree [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1258.763806] env[62753]: INFO nova.compute.manager [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Detaching volume 95aec84e-bc3f-4074-b379-2b5bf5976d96 [ 1258.799740] env[62753]: INFO nova.virt.block_device [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Attempting to driver detach volume 95aec84e-bc3f-4074-b379-2b5bf5976d96 from mountpoint /dev/sdb [ 1258.800096] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Volume detach. Driver type: vmdk {{(pid=62753) detach_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:646}} [ 1258.800337] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] _detach_volume_vmdk: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284708', 'volume_id': '95aec84e-bc3f-4074-b379-2b5bf5976d96', 'name': 'volume-95aec84e-bc3f-4074-b379-2b5bf5976d96', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '221f76f0-ee69-4119-b19b-c148e40c1126', 'attached_at': '', 'detached_at': '', 'volume_id': '95aec84e-bc3f-4074-b379-2b5bf5976d96', 'serial': '95aec84e-bc3f-4074-b379-2b5bf5976d96'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:571}} [ 1258.801329] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5ca66f81-86aa-48c4-ba96-40874147bfd9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.830799] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fd22efa0-0b1b-4d4f-be7a-50f617ff866f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.839653] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-65ac33b9-1637-4f2a-b573-d678d78066bf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.866090] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0470fdaf-3188-45f8-8697-2b36305c954a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.881439] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] The volume has not been displaced from its original location: [datastore2] volume-95aec84e-bc3f-4074-b379-2b5bf5976d96/volume-95aec84e-bc3f-4074-b379-2b5bf5976d96.vmdk. No consolidation needed. {{(pid=62753) _consolidate_vmdk_volume /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:504}} [ 1258.886743] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Reconfiguring VM instance instance-0000006f to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1258.891648] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-5a256f73-1755-4e13-88b1-66360fd10ca2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1258.912564] env[62753]: DEBUG oslo_vmware.api [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1258.912564] env[62753]: value = "task-1332992" [ 1258.912564] env[62753]: _type = "Task" [ 1258.912564] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1258.922582] env[62753]: DEBUG oslo_vmware.api [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332992, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.178930] env[62753]: DEBUG nova.scheduler.client.report [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1259.212890] env[62753]: DEBUG oslo_vmware.rw_handles [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Completed reading data from the image iterator. {{(pid=62753) read /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:765}} [ 1259.213134] env[62753]: DEBUG oslo_vmware.rw_handles [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Getting lease state for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52128711-3166-1e28-e12c-541efbcac951/disk-0.vmdk. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:423}} [ 1259.214841] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cd8ba7b-7bce-4198-987f-55a7abf98ef2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.221892] env[62753]: DEBUG oslo_vmware.rw_handles [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52128711-3166-1e28-e12c-541efbcac951/disk-0.vmdk is in state: ready. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:430}} [ 1259.222186] env[62753]: DEBUG oslo_vmware.rw_handles [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lease for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52128711-3166-1e28-e12c-541efbcac951/disk-0.vmdk. {{(pid=62753) _release_lease /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:440}} [ 1259.222534] env[62753]: DEBUG oslo_vmware.service [-] Invoking HttpNfcLease.HttpNfcLeaseComplete with opID=oslo.vmware-031c7c0a-48bd-459e-9a94-0511b727ca3e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.422618] env[62753]: DEBUG oslo_vmware.api [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332992, 'name': ReconfigVM_Task, 'duration_secs': 0.235923} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.424080] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Reconfigured VM instance instance-0000006f to detach disk 2001 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1259.427967] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-cffc0fa4-f7e6-47f4-bdb4-f1e64a0bc9a5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.442996] env[62753]: DEBUG oslo_vmware.api [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1259.442996] env[62753]: value = "task-1332993" [ 1259.442996] env[62753]: _type = "Task" [ 1259.442996] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.451168] env[62753]: DEBUG oslo_vmware.api [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332993, 'name': ReconfigVM_Task} progress is 6%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.556897] env[62753]: DEBUG oslo_vmware.rw_handles [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Closed VMDK write handle for https://esx7c1n2.openstack.eu-de-1.cloud.sap/nfc/52128711-3166-1e28-e12c-541efbcac951/disk-0.vmdk. {{(pid=62753) close /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/rw_handles.py:646}} [ 1259.557211] env[62753]: INFO nova.virt.vmwareapi.images [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Downloaded image file data 026f5cd9-1845-4b09-9364-71568c7d3e7f [ 1259.558077] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7a26086c-ab94-419d-b8f5-4d0250bb2dbf {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.572921] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a293722a-db0e-457c-9328-848641a600ca {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.624347] env[62753]: INFO nova.virt.vmwareapi.images [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] The imported VM was unregistered [ 1259.626701] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Caching image {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:638}} [ 1259.627017] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Creating directory with path [datastore2] devstack-image-cache_base/026f5cd9-1845-4b09-9364-71568c7d3e7f {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1259.627298] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c51d9e32-0701-438b-978f-01f2039239fb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.636658] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Created directory with path [datastore2] devstack-image-cache_base/026f5cd9-1845-4b09-9364-71568c7d3e7f {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1259.636891] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Moving virtual disk from [datastore2] OSTACK_IMG_f876bfa5-fcb8-4250-a141-f794a40f5863/OSTACK_IMG_f876bfa5-fcb8-4250-a141-f794a40f5863.vmdk to [datastore2] devstack-image-cache_base/026f5cd9-1845-4b09-9364-71568c7d3e7f/026f5cd9-1845-4b09-9364-71568c7d3e7f.vmdk. {{(pid=62753) disk_move /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:263}} [ 1259.637173] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.MoveVirtualDisk_Task with opID=oslo.vmware-6e8c3767-5fec-4ab3-88d8-e5b4b89f8e32 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1259.642916] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1259.642916] env[62753]: value = "task-1332995" [ 1259.642916] env[62753]: _type = "Task" [ 1259.642916] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1259.650272] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332995, 'name': MoveVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1259.952477] env[62753]: DEBUG oslo_vmware.api [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1332993, 'name': ReconfigVM_Task, 'duration_secs': 0.158092} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1259.952797] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Detached VMDK: {'driver_volume_type': 'vmdk', 'data': {'volume': 'vm-284708', 'volume_id': '95aec84e-bc3f-4074-b379-2b5bf5976d96', 'name': 'volume-95aec84e-bc3f-4074-b379-2b5bf5976d96', 'profile_id': None, 'qos_specs': None, 'access_mode': 'rw', 'encrypted': False, 'cacheable': False}, 'status': 'reserved', 'instance': '221f76f0-ee69-4119-b19b-c148e40c1126', 'attached_at': '', 'detached_at': '', 'volume_id': '95aec84e-bc3f-4074-b379-2b5bf5976d96', 'serial': '95aec84e-bc3f-4074-b379-2b5bf5976d96'} {{(pid=62753) _detach_volume_vmdk /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:605}} [ 1260.152838] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332995, 'name': MoveVirtualDisk_Task} progress is 24%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.190436] env[62753]: DEBUG oslo_concurrency.lockutils [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_source" :: held 2.163s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1260.497124] env[62753]: DEBUG nova.objects.instance [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lazy-loading 'flavor' on Instance uuid 221f76f0-ee69-4119-b19b-c148e40c1126 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1260.652491] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332995, 'name': MoveVirtualDisk_Task} progress is 49%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1260.752234] env[62753]: INFO nova.scheduler.client.report [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Deleted allocation for migration 1491faad-0fb0-4f1b-8bed-510b0da4f0c1 [ 1260.843581] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Didn't find any instances for network info cache update. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1260.843581] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1261.154708] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332995, 'name': MoveVirtualDisk_Task} progress is 71%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.254644] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "22cea165-cfde-403f-a0e2-82a86f5baa51" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1261.257980] env[62753]: DEBUG oslo_concurrency.lockutils [None req-649c904d-ab7e-45e9-a01a-5446e68f7fa2 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51" "released" by "nova.compute.manager.ComputeManager.confirm_resize..do_confirm_resize" :: held 6.148s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1261.259185] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.005s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1261.259458] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "22cea165-cfde-403f-a0e2-82a86f5baa51-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1261.259698] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1261.259874] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1261.261824] env[62753]: INFO nova.compute.manager [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Terminating instance [ 1261.263609] env[62753]: DEBUG nova.compute.manager [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1261.263831] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1261.264702] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4302b146-4add-46e3-9e2f-22a6bc9e263e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.272785] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1261.273595] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b6109903-83d0-402a-8420-fc52769ef568 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.283598] env[62753]: DEBUG oslo_vmware.api [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1261.283598] env[62753]: value = "task-1332996" [ 1261.283598] env[62753]: _type = "Task" [ 1261.283598] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.292118] env[62753]: DEBUG oslo_vmware.api [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332996, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.506758] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2d38593d-8d60-4a2d-9d1f-89f04cc785d5 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "221f76f0-ee69-4119-b19b-c148e40c1126" "released" by "nova.compute.manager.ComputeManager.detach_volume..do_detach_volume" :: held 3.246s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1261.653526] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332995, 'name': MoveVirtualDisk_Task} progress is 94%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1261.794702] env[62753]: DEBUG oslo_vmware.api [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332996, 'name': PowerOffVM_Task, 'duration_secs': 0.297616} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1261.794926] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1261.795116] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1261.795367] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-e1bc90aa-5908-48a1-bbb8-8bd938557fcb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.855218] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1261.855437] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1261.855625] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Deleting the datastore file [datastore1] 22cea165-cfde-403f-a0e2-82a86f5baa51 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1261.855896] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-336ff524-843b-4cd5-92ac-8f2cd386a9f8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1261.861753] env[62753]: DEBUG oslo_vmware.api [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1261.861753] env[62753]: value = "task-1332998" [ 1261.861753] env[62753]: _type = "Task" [ 1261.861753] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1261.869903] env[62753]: DEBUG oslo_vmware.api [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332998, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.153935] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332995, 'name': MoveVirtualDisk_Task, 'duration_secs': 2.226005} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.154230] env[62753]: INFO nova.virt.vmwareapi.ds_util [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Moved virtual disk from [datastore2] OSTACK_IMG_f876bfa5-fcb8-4250-a141-f794a40f5863/OSTACK_IMG_f876bfa5-fcb8-4250-a141-f794a40f5863.vmdk to [datastore2] devstack-image-cache_base/026f5cd9-1845-4b09-9364-71568c7d3e7f/026f5cd9-1845-4b09-9364-71568c7d3e7f.vmdk. [ 1262.154426] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Cleaning up location [datastore2] OSTACK_IMG_f876bfa5-fcb8-4250-a141-f794a40f5863 {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:640}} [ 1262.154592] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Deleting the datastore file [datastore2] OSTACK_IMG_f876bfa5-fcb8-4250-a141-f794a40f5863 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1262.154853] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-ba889bfb-4d47-4d19-ba58-519607e01a2d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.161242] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1262.161242] env[62753]: value = "task-1332999" [ 1262.161242] env[62753]: _type = "Task" [ 1262.161242] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.168960] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332999, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.371203] env[62753]: DEBUG oslo_vmware.api [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1332998, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.182208} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.371473] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1262.371665] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1262.371842] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1262.372037] env[62753]: INFO nova.compute.manager [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1262.372300] env[62753]: DEBUG oslo.service.loopingcall [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1262.372495] env[62753]: DEBUG nova.compute.manager [-] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1262.372589] env[62753]: DEBUG nova.network.neutron [-] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1262.509861] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "221f76f0-ee69-4119-b19b-c148e40c1126" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1262.510195] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "221f76f0-ee69-4119-b19b-c148e40c1126" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1262.510398] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "221f76f0-ee69-4119-b19b-c148e40c1126-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1262.510585] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "221f76f0-ee69-4119-b19b-c148e40c1126-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1262.510747] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "221f76f0-ee69-4119-b19b-c148e40c1126-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1262.512885] env[62753]: INFO nova.compute.manager [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Terminating instance [ 1262.514563] env[62753]: DEBUG nova.compute.manager [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1262.514759] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1262.515629] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3fb820fa-2587-4d14-8896-f9d3e6dcf94b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.522943] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1262.523180] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-32299cb9-9b80-4ef0-99b2-e64552f57769 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.528513] env[62753]: DEBUG oslo_vmware.api [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1262.528513] env[62753]: value = "task-1333000" [ 1262.528513] env[62753]: _type = "Task" [ 1262.528513] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.535804] env[62753]: DEBUG oslo_vmware.api [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1333000, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.671416] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1332999, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.033918} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1262.671669] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1262.671988] env[62753]: DEBUG oslo_concurrency.lockutils [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Releasing lock "[datastore2] devstack-image-cache_base/026f5cd9-1845-4b09-9364-71568c7d3e7f/026f5cd9-1845-4b09-9364-71568c7d3e7f.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1262.672100] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/026f5cd9-1845-4b09-9364-71568c7d3e7f/026f5cd9-1845-4b09-9364-71568c7d3e7f.vmdk to [datastore2] 2e784985-7045-448e-a650-f6692bb69d94/2e784985-7045-448e-a650-f6692bb69d94.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1262.672367] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-ec53470d-f8f2-4cfb-97a2-f2f3b86fe141 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1262.680594] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1262.680594] env[62753]: value = "task-1333001" [ 1262.680594] env[62753]: _type = "Task" [ 1262.680594] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1262.688454] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333001, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1262.814496] env[62753]: DEBUG nova.compute.manager [req-4e7487c6-064b-4301-a951-e156f79dd6c7 req-ff07780f-63fc-4b4b-b905-1acaeb68e05a service nova] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Received event network-vif-deleted-25e4c4c8-9d78-4a88-9a5e-4a364ee574a5 {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1262.814679] env[62753]: INFO nova.compute.manager [req-4e7487c6-064b-4301-a951-e156f79dd6c7 req-ff07780f-63fc-4b4b-b905-1acaeb68e05a service nova] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Neutron deleted interface 25e4c4c8-9d78-4a88-9a5e-4a364ee574a5; detaching it from the instance and deleting it from the info cache [ 1262.815092] env[62753]: DEBUG nova.network.neutron [req-4e7487c6-064b-4301-a951-e156f79dd6c7 req-ff07780f-63fc-4b4b-b905-1acaeb68e05a service nova] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1263.039110] env[62753]: DEBUG oslo_vmware.api [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1333000, 'name': PowerOffVM_Task, 'duration_secs': 0.227012} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.039604] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1263.039802] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1263.040111] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a6be3271-a22e-41f6-8169-b553e17a6610 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.104573] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1263.104820] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1263.105119] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Deleting the datastore file [datastore1] 221f76f0-ee69-4119-b19b-c148e40c1126 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1263.105399] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-bde00c57-ade9-4f08-804f-711052e30ba6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.111788] env[62753]: DEBUG oslo_vmware.api [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for the task: (returnval){ [ 1263.111788] env[62753]: value = "task-1333003" [ 1263.111788] env[62753]: _type = "Task" [ 1263.111788] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1263.121856] env[62753]: DEBUG oslo_vmware.api [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1333003, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.191053] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333001, 'name': CopyVirtualDisk_Task} progress is 15%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.293514] env[62753]: DEBUG nova.network.neutron [-] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1263.317724] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-dda01bb8-b795-442a-b537-f453f2cd5211 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.327170] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-709d59e3-ed18-4217-9679-bd1b884bad70 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1263.351936] env[62753]: DEBUG nova.compute.manager [req-4e7487c6-064b-4301-a951-e156f79dd6c7 req-ff07780f-63fc-4b4b-b905-1acaeb68e05a service nova] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Detach interface failed, port_id=25e4c4c8-9d78-4a88-9a5e-4a364ee574a5, reason: Instance 22cea165-cfde-403f-a0e2-82a86f5baa51 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1263.621408] env[62753]: DEBUG oslo_vmware.api [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Task: {'id': task-1333003, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.264485} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1263.621731] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1263.621853] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1263.622045] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1263.622267] env[62753]: INFO nova.compute.manager [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1263.622502] env[62753]: DEBUG oslo.service.loopingcall [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1263.622738] env[62753]: DEBUG nova.compute.manager [-] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1263.622825] env[62753]: DEBUG nova.network.neutron [-] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1263.692822] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333001, 'name': CopyVirtualDisk_Task} progress is 35%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1263.796685] env[62753]: INFO nova.compute.manager [-] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Took 1.42 seconds to deallocate network for instance. [ 1264.192371] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333001, 'name': CopyVirtualDisk_Task} progress is 57%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.304783] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1264.304783] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1264.304987] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1264.328128] env[62753]: INFO nova.scheduler.client.report [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Deleted allocations for instance 22cea165-cfde-403f-a0e2-82a86f5baa51 [ 1264.692653] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333001, 'name': CopyVirtualDisk_Task} progress is 80%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1264.785887] env[62753]: DEBUG nova.network.neutron [-] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1264.838879] env[62753]: DEBUG oslo_concurrency.lockutils [None req-9f207eb4-bfac-42f3-98f0-e3bd80d66354 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "22cea165-cfde-403f-a0e2-82a86f5baa51" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.579s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1264.841827] env[62753]: DEBUG nova.compute.manager [req-ce30eef9-9e06-4128-a647-6ef2a166a7ef req-a68568f7-6812-43ee-894c-2a412a13256b service nova] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Received event network-vif-deleted-7d3a7c1e-e1c1-42da-aacb-dad06fa6049a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1265.192662] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333001, 'name': CopyVirtualDisk_Task, 'duration_secs': 2.415831} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1265.192778] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/026f5cd9-1845-4b09-9364-71568c7d3e7f/026f5cd9-1845-4b09-9364-71568c7d3e7f.vmdk to [datastore2] 2e784985-7045-448e-a650-f6692bb69d94/2e784985-7045-448e-a650-f6692bb69d94.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1265.193969] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a1e33b7f-dc30-4b54-8421-690a3906d2c2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.216083] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Reconfiguring VM instance instance-0000006e to attach disk [datastore2] 2e784985-7045-448e-a650-f6692bb69d94/2e784985-7045-448e-a650-f6692bb69d94.vmdk or device None with type streamOptimized {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1265.216237] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-04e86dc7-5bef-4688-8f7f-860b54c0d2b5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1265.235542] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1265.235542] env[62753]: value = "task-1333004" [ 1265.235542] env[62753]: _type = "Task" [ 1265.235542] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1265.243457] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333004, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.289109] env[62753]: INFO nova.compute.manager [-] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Took 1.67 seconds to deallocate network for instance. [ 1265.745931] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333004, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1265.795132] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1265.795404] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1265.795628] env[62753]: DEBUG nova.objects.instance [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lazy-loading 'resources' on Instance uuid 221f76f0-ee69-4119-b19b-c148e40c1126 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1266.212081] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "5b196309-161a-4770-90c8-81b57909cb12" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1266.212341] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "5b196309-161a-4770-90c8-81b57909cb12" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1266.248832] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333004, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.387321] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b9f3c680-1bba-4378-a35f-2cd4ee9bb47c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.394840] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6432d1a7-1087-4b28-ad2b-7faa7a1257cc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.426152] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-39b41f7d-b953-4648-824e-74fa8f903e61 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.433418] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d6186452-3cc4-49f1-b433-f6971ffd7399 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.446222] env[62753]: DEBUG nova.compute.provider_tree [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1266.714739] env[62753]: DEBUG nova.compute.manager [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1266.746295] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333004, 'name': ReconfigVM_Task, 'duration_secs': 1.027913} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1266.746613] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Reconfigured VM instance instance-0000006e to attach disk [datastore2] 2e784985-7045-448e-a650-f6692bb69d94/2e784985-7045-448e-a650-f6692bb69d94.vmdk or device None with type streamOptimized {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1266.747274] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-4a64cf5b-a028-4a5d-b518-8b03d04ec207 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1266.753653] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1266.753653] env[62753]: value = "task-1333005" [ 1266.753653] env[62753]: _type = "Task" [ 1266.753653] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1266.760934] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333005, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1266.949620] env[62753]: DEBUG nova.scheduler.client.report [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1267.235379] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1267.263483] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333005, 'name': Rename_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.454745] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.659s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1267.457872] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.222s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1267.458874] env[62753]: INFO nova.compute.claims [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1267.475081] env[62753]: INFO nova.scheduler.client.report [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Deleted allocations for instance 221f76f0-ee69-4119-b19b-c148e40c1126 [ 1267.764041] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333005, 'name': Rename_Task, 'duration_secs': 0.55905} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1267.764424] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1267.764636] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-280c2e0f-6a46-43db-9e92-45f825946390 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1267.770410] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1267.770410] env[62753]: value = "task-1333006" [ 1267.770410] env[62753]: _type = "Task" [ 1267.770410] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1267.780190] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333006, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1267.982767] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4b456fbb-be68-432b-b0ad-2da44a8296b0 tempest-AttachVolumeNegativeTest-421529539 tempest-AttachVolumeNegativeTest-421529539-project-member] Lock "221f76f0-ee69-4119-b19b-c148e40c1126" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.472s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1268.280736] env[62753]: DEBUG oslo_vmware.api [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333006, 'name': PowerOnVM_Task, 'duration_secs': 0.45838} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1268.281047] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1268.382094] env[62753]: DEBUG nova.compute.manager [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1268.383023] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7ae4a6a5-716a-47c8-9f62-a107dd17f02d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.502363] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-74099618-48e3-4ed4-a813-f15467d82576 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.509543] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f8251a02-3eb7-4b88-85cd-46ce8fb2e0d0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.538845] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c20113b-d905-48c0-86c8-ca996e786b6a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.546509] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1efe6442-8214-42c2-a177-7d8a9eec3e8c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1268.560348] env[62753]: DEBUG nova.compute.provider_tree [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1268.905013] env[62753]: DEBUG oslo_concurrency.lockutils [None req-26dfcd9e-ca86-47b6-b0d3-0c4f2529aaf5 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "2e784985-7045-448e-a650-f6692bb69d94" "released" by "nova.compute.manager.ComputeManager.unshelve_instance..do_unshelve_instance" :: held 20.842s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1269.063347] env[62753]: DEBUG nova.scheduler.client.report [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1269.567901] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.111s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1269.568542] env[62753]: DEBUG nova.compute.manager [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1269.596759] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "2e784985-7045-448e-a650-f6692bb69d94" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1269.596759] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "2e784985-7045-448e-a650-f6692bb69d94" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.596759] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "2e784985-7045-448e-a650-f6692bb69d94-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1269.596996] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "2e784985-7045-448e-a650-f6692bb69d94-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1269.596996] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "2e784985-7045-448e-a650-f6692bb69d94-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1269.599147] env[62753]: INFO nova.compute.manager [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Terminating instance [ 1269.600927] env[62753]: DEBUG nova.compute.manager [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1269.601138] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1269.601956] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6697b26b-f1f9-4fd3-8bff-2fc1e238196b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.610402] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1269.610622] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-f7aafdbd-d84e-4b2c-a19a-a0d38ef22a70 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1269.616434] env[62753]: DEBUG oslo_vmware.api [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1269.616434] env[62753]: value = "task-1333008" [ 1269.616434] env[62753]: _type = "Task" [ 1269.616434] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1269.625545] env[62753]: DEBUG oslo_vmware.api [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333008, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.073630] env[62753]: DEBUG nova.compute.utils [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1270.075068] env[62753]: DEBUG nova.compute.manager [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1270.075267] env[62753]: DEBUG nova.network.neutron [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1270.121540] env[62753]: DEBUG nova.policy [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd65bf0df728640ef9104e14d6320ed01', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '891d1efbcac34ceba5fa72ea6b5fe38b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1270.128432] env[62753]: DEBUG oslo_vmware.api [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333008, 'name': PowerOffVM_Task, 'duration_secs': 0.190869} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.128672] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1270.128844] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1270.129102] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-78831dfa-3dd3-4ef6-b387-1939d226d114 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.190662] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1270.190888] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1270.191087] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Deleting the datastore file [datastore2] 2e784985-7045-448e-a650-f6692bb69d94 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1270.191360] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-b5a9b374-3c22-4a1d-b080-b495e24b6d52 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1270.197854] env[62753]: DEBUG oslo_vmware.api [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for the task: (returnval){ [ 1270.197854] env[62753]: value = "task-1333010" [ 1270.197854] env[62753]: _type = "Task" [ 1270.197854] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1270.206123] env[62753]: DEBUG oslo_vmware.api [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333010, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1270.391576] env[62753]: DEBUG nova.network.neutron [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Successfully created port: 33f767b7-a237-433b-9476-7cb32f4af66a {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1270.578711] env[62753]: DEBUG nova.compute.manager [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1270.707874] env[62753]: DEBUG oslo_vmware.api [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Task: {'id': task-1333010, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.157158} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1270.708166] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1270.708370] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1270.708551] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1270.708728] env[62753]: INFO nova.compute.manager [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1270.708980] env[62753]: DEBUG oslo.service.loopingcall [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1270.709197] env[62753]: DEBUG nova.compute.manager [-] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1270.709296] env[62753]: DEBUG nova.network.neutron [-] [instance: 2e784985-7045-448e-a650-f6692bb69d94] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1271.179543] env[62753]: DEBUG nova.compute.manager [req-8385f148-f8c5-4b19-8a9b-9589e595f04b req-abcf01bf-13b5-495c-baa7-822c461f460a service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Received event network-vif-deleted-b81e545c-6410-4acd-8777-1c9a2ef2b7eb {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1271.179796] env[62753]: INFO nova.compute.manager [req-8385f148-f8c5-4b19-8a9b-9589e595f04b req-abcf01bf-13b5-495c-baa7-822c461f460a service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Neutron deleted interface b81e545c-6410-4acd-8777-1c9a2ef2b7eb; detaching it from the instance and deleting it from the info cache [ 1271.179986] env[62753]: DEBUG nova.network.neutron [req-8385f148-f8c5-4b19-8a9b-9589e595f04b req-abcf01bf-13b5-495c-baa7-822c461f460a service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1271.592762] env[62753]: DEBUG nova.compute.manager [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1271.617645] env[62753]: DEBUG nova.virt.hardware [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1271.617893] env[62753]: DEBUG nova.virt.hardware [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1271.618087] env[62753]: DEBUG nova.virt.hardware [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1271.618339] env[62753]: DEBUG nova.virt.hardware [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1271.618501] env[62753]: DEBUG nova.virt.hardware [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1271.618659] env[62753]: DEBUG nova.virt.hardware [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1271.618874] env[62753]: DEBUG nova.virt.hardware [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1271.619055] env[62753]: DEBUG nova.virt.hardware [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1271.619237] env[62753]: DEBUG nova.virt.hardware [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1271.619411] env[62753]: DEBUG nova.virt.hardware [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1271.619585] env[62753]: DEBUG nova.virt.hardware [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1271.620461] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9396017b-fc27-4074-8e4c-ed906c300fcd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.629415] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-eaf89744-57a6-4d45-8a7b-f89e8d0efecb {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.653675] env[62753]: DEBUG nova.network.neutron [-] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1271.684028] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-535a7f70-7aa8-4db8-be7b-c5286d228a5f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.692473] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e2a94a43-fc78-46a8-a021-47822ee8f9d3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1271.715405] env[62753]: DEBUG nova.compute.manager [req-8385f148-f8c5-4b19-8a9b-9589e595f04b req-abcf01bf-13b5-495c-baa7-822c461f460a service nova] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Detach interface failed, port_id=b81e545c-6410-4acd-8777-1c9a2ef2b7eb, reason: Instance 2e784985-7045-448e-a650-f6692bb69d94 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1271.851184] env[62753]: DEBUG nova.network.neutron [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Successfully updated port: 33f767b7-a237-433b-9476-7cb32f4af66a {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1272.156061] env[62753]: INFO nova.compute.manager [-] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Took 1.45 seconds to deallocate network for instance. [ 1272.354115] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1272.354383] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1272.354534] env[62753]: DEBUG nova.network.neutron [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1272.664437] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1272.664437] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1272.664437] env[62753]: DEBUG nova.objects.instance [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lazy-loading 'resources' on Instance uuid 2e784985-7045-448e-a650-f6692bb69d94 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1272.885155] env[62753]: DEBUG nova.network.neutron [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1273.011463] env[62753]: DEBUG nova.network.neutron [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance_info_cache with network_info: [{"id": "33f767b7-a237-433b-9476-7cb32f4af66a", "address": "fa:16:3e:d0:d1:9b", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33f767b7-a2", "ovs_interfaceid": "33f767b7-a237-433b-9476-7cb32f4af66a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1273.204246] env[62753]: DEBUG nova.compute.manager [req-a41bc277-677e-4e56-bd08-c27771b8d512 req-0a43f9cc-86ae-441d-95fd-01d861f897b7 service nova] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Received event network-vif-plugged-33f767b7-a237-433b-9476-7cb32f4af66a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1273.204449] env[62753]: DEBUG oslo_concurrency.lockutils [req-a41bc277-677e-4e56-bd08-c27771b8d512 req-0a43f9cc-86ae-441d-95fd-01d861f897b7 service nova] Acquiring lock "5b196309-161a-4770-90c8-81b57909cb12-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1273.204646] env[62753]: DEBUG oslo_concurrency.lockutils [req-a41bc277-677e-4e56-bd08-c27771b8d512 req-0a43f9cc-86ae-441d-95fd-01d861f897b7 service nova] Lock "5b196309-161a-4770-90c8-81b57909cb12-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1273.204818] env[62753]: DEBUG oslo_concurrency.lockutils [req-a41bc277-677e-4e56-bd08-c27771b8d512 req-0a43f9cc-86ae-441d-95fd-01d861f897b7 service nova] Lock "5b196309-161a-4770-90c8-81b57909cb12-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1273.205068] env[62753]: DEBUG nova.compute.manager [req-a41bc277-677e-4e56-bd08-c27771b8d512 req-0a43f9cc-86ae-441d-95fd-01d861f897b7 service nova] [instance: 5b196309-161a-4770-90c8-81b57909cb12] No waiting events found dispatching network-vif-plugged-33f767b7-a237-433b-9476-7cb32f4af66a {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1273.205168] env[62753]: WARNING nova.compute.manager [req-a41bc277-677e-4e56-bd08-c27771b8d512 req-0a43f9cc-86ae-441d-95fd-01d861f897b7 service nova] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Received unexpected event network-vif-plugged-33f767b7-a237-433b-9476-7cb32f4af66a for instance with vm_state building and task_state spawning. [ 1273.205391] env[62753]: DEBUG nova.compute.manager [req-a41bc277-677e-4e56-bd08-c27771b8d512 req-0a43f9cc-86ae-441d-95fd-01d861f897b7 service nova] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Received event network-changed-33f767b7-a237-433b-9476-7cb32f4af66a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1273.205577] env[62753]: DEBUG nova.compute.manager [req-a41bc277-677e-4e56-bd08-c27771b8d512 req-0a43f9cc-86ae-441d-95fd-01d861f897b7 service nova] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Refreshing instance network info cache due to event network-changed-33f767b7-a237-433b-9476-7cb32f4af66a. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1273.205759] env[62753]: DEBUG oslo_concurrency.lockutils [req-a41bc277-677e-4e56-bd08-c27771b8d512 req-0a43f9cc-86ae-441d-95fd-01d861f897b7 service nova] Acquiring lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1273.209117] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22de88ef-07ea-4110-a375-0709375eaba5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.217036] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f906c247-cade-4b96-b743-e81c6b5ef80a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.246890] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6b0a8a96-fc1b-4d5e-b2ac-565a6f43d4f2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.254972] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f3f8a213-c158-4d5c-b0c5-02204b6ffb80 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.267851] env[62753]: DEBUG nova.compute.provider_tree [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1273.514086] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1273.514439] env[62753]: DEBUG nova.compute.manager [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Instance network_info: |[{"id": "33f767b7-a237-433b-9476-7cb32f4af66a", "address": "fa:16:3e:d0:d1:9b", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33f767b7-a2", "ovs_interfaceid": "33f767b7-a237-433b-9476-7cb32f4af66a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1273.514769] env[62753]: DEBUG oslo_concurrency.lockutils [req-a41bc277-677e-4e56-bd08-c27771b8d512 req-0a43f9cc-86ae-441d-95fd-01d861f897b7 service nova] Acquired lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1273.514959] env[62753]: DEBUG nova.network.neutron [req-a41bc277-677e-4e56-bd08-c27771b8d512 req-0a43f9cc-86ae-441d-95fd-01d861f897b7 service nova] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Refreshing network info cache for port 33f767b7-a237-433b-9476-7cb32f4af66a {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1273.516257] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:d0:d1:9b', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cae1d6a8-cbba-4bbf-af10-ba5467340475', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '33f767b7-a237-433b-9476-7cb32f4af66a', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1273.523866] env[62753]: DEBUG oslo.service.loopingcall [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1273.526552] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1273.527043] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-d1a01c28-8fd7-438b-a996-194da10091ae {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1273.546423] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1273.546423] env[62753]: value = "task-1333012" [ 1273.546423] env[62753]: _type = "Task" [ 1273.546423] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1273.553557] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333012, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1273.725683] env[62753]: DEBUG nova.network.neutron [req-a41bc277-677e-4e56-bd08-c27771b8d512 req-0a43f9cc-86ae-441d-95fd-01d861f897b7 service nova] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updated VIF entry in instance network info cache for port 33f767b7-a237-433b-9476-7cb32f4af66a. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1273.726136] env[62753]: DEBUG nova.network.neutron [req-a41bc277-677e-4e56-bd08-c27771b8d512 req-0a43f9cc-86ae-441d-95fd-01d861f897b7 service nova] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance_info_cache with network_info: [{"id": "33f767b7-a237-433b-9476-7cb32f4af66a", "address": "fa:16:3e:d0:d1:9b", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33f767b7-a2", "ovs_interfaceid": "33f767b7-a237-433b-9476-7cb32f4af66a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1273.771025] env[62753]: DEBUG nova.scheduler.client.report [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1274.056720] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333012, 'name': CreateVM_Task, 'duration_secs': 0.3056} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.056881] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1274.057626] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1274.057794] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1274.058183] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired external semaphore "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1274.058420] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-a9407907-0961-476c-b5d8-425cde8ffcb1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.063122] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1274.063122] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bc81f8-58e6-4f94-877c-858d679c3239" [ 1274.063122] env[62753]: _type = "Task" [ 1274.063122] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.070744] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bc81f8-58e6-4f94-877c-858d679c3239, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.228750] env[62753]: DEBUG oslo_concurrency.lockutils [req-a41bc277-677e-4e56-bd08-c27771b8d512 req-0a43f9cc-86ae-441d-95fd-01d861f897b7 service nova] Releasing lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1274.276565] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.613s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1274.293950] env[62753]: INFO nova.scheduler.client.report [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Deleted allocations for instance 2e784985-7045-448e-a650-f6692bb69d94 [ 1274.575343] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bc81f8-58e6-4f94-877c-858d679c3239, 'name': SearchDatastore_Task, 'duration_secs': 0.009234} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1274.575694] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1274.575983] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1274.576296] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1274.576483] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1274.576673] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Creating directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1274.576955] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-c4a2bf52-66b0-4c08-9b9d-5d79a13bbcb5 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.583852] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Created directory with path [datastore1] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1274.584069] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Folder [datastore1] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1274.584826] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-65085075-9ffe-4803-b040-b28890118c5d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1274.589851] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1274.589851] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a05b64-2c66-69dd-a15c-7d948adc19f0" [ 1274.589851] env[62753]: _type = "Task" [ 1274.589851] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1274.597311] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a05b64-2c66-69dd-a15c-7d948adc19f0, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1274.801797] env[62753]: DEBUG oslo_concurrency.lockutils [None req-4e2f1592-8e8c-4940-be34-f6ea92d7e648 tempest-ServerActionsTestOtherB-1429955500 tempest-ServerActionsTestOtherB-1429955500-project-member] Lock "2e784985-7045-448e-a650-f6692bb69d94" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.205s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1275.100475] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52a05b64-2c66-69dd-a15c-7d948adc19f0, 'name': SearchDatastore_Task, 'duration_secs': 0.007412} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1275.149438] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-27d65adb-d328-4009-a377-b1ae41590e28 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.149438] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1275.149438] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52635664-bdc5-e17e-0444-dbf0250f1f43" [ 1275.149438] env[62753]: _type = "Task" [ 1275.149438] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1275.149438] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52635664-bdc5-e17e-0444-dbf0250f1f43, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1275.618116] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52635664-bdc5-e17e-0444-dbf0250f1f43, 'name': SearchDatastore_Task, 'duration_secs': 0.017488} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1275.618456] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "[datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1275.618583] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Copying Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 5b196309-161a-4770-90c8-81b57909cb12/5b196309-161a-4770-90c8-81b57909cb12.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1275.618833] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-d302dbfb-f908-4c43-bf4e-05bc0ef0cdd7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1275.624889] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1275.624889] env[62753]: value = "task-1333015" [ 1275.624889] env[62753]: _type = "Task" [ 1275.624889] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1275.631868] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333015, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.134531] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333015, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.477258} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1276.134786] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Copied Virtual Disk [datastore1] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore1] 5b196309-161a-4770-90c8-81b57909cb12/5b196309-161a-4770-90c8-81b57909cb12.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1276.135014] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1276.135271] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-b066176a-d64e-48e2-bd97-ae60d22eaa4a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.141985] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1276.141985] env[62753]: value = "task-1333016" [ 1276.141985] env[62753]: _type = "Task" [ 1276.141985] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.149930] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333016, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1276.651815] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333016, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.064613} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1276.652113] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1276.652906] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8c18e9a2-1557-48a8-aa9d-842984002bd7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.673985] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] 5b196309-161a-4770-90c8-81b57909cb12/5b196309-161a-4770-90c8-81b57909cb12.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1276.674264] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-797884ac-ecfd-4247-b5bb-797ae4827a53 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1276.693364] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1276.693364] env[62753]: value = "task-1333017" [ 1276.693364] env[62753]: _type = "Task" [ 1276.693364] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1276.700634] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333017, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.203514] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333017, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1277.707152] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333017, 'name': ReconfigVM_Task} progress is 14%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.204314] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333017, 'name': ReconfigVM_Task, 'duration_secs': 1.431601} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.204763] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Reconfigured VM instance instance-00000071 to attach disk [datastore1] 5b196309-161a-4770-90c8-81b57909cb12/5b196309-161a-4770-90c8-81b57909cb12.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1278.205469] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-93b51c3b-d860-4222-9b50-a30b883b805d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.211250] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1278.211250] env[62753]: value = "task-1333018" [ 1278.211250] env[62753]: _type = "Task" [ 1278.211250] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1278.219039] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333018, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1278.722135] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333018, 'name': Rename_Task, 'duration_secs': 0.133167} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1278.722135] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1278.722447] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-8a553e5a-4cdd-44b3-92a9-ed379003ecd1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1278.728912] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1278.728912] env[62753]: value = "task-1333019" [ 1278.728912] env[62753]: _type = "Task" [ 1278.728912] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1278.737215] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333019, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1279.239223] env[62753]: DEBUG oslo_vmware.api [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333019, 'name': PowerOnVM_Task, 'duration_secs': 0.437309} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1279.239537] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1279.239766] env[62753]: INFO nova.compute.manager [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Took 7.65 seconds to spawn the instance on the hypervisor. [ 1279.239987] env[62753]: DEBUG nova.compute.manager [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1279.241169] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-b95567cd-56cf-4484-8624-eac59dd8c6fc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1279.758750] env[62753]: INFO nova.compute.manager [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Took 12.54 seconds to build instance. [ 1280.260313] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c1e19d6-bd2f-4fc8-ab6d-ee6429edfcfe tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "5b196309-161a-4770-90c8-81b57909cb12" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 14.048s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1280.409286] env[62753]: DEBUG nova.compute.manager [req-fd958215-b163-4502-98cf-c291899fc22c req-cedf4fed-1609-4049-86aa-3762492472d5 service nova] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Received event network-changed-33f767b7-a237-433b-9476-7cb32f4af66a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1280.409536] env[62753]: DEBUG nova.compute.manager [req-fd958215-b163-4502-98cf-c291899fc22c req-cedf4fed-1609-4049-86aa-3762492472d5 service nova] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Refreshing instance network info cache due to event network-changed-33f767b7-a237-433b-9476-7cb32f4af66a. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1280.409838] env[62753]: DEBUG oslo_concurrency.lockutils [req-fd958215-b163-4502-98cf-c291899fc22c req-cedf4fed-1609-4049-86aa-3762492472d5 service nova] Acquiring lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1280.409990] env[62753]: DEBUG oslo_concurrency.lockutils [req-fd958215-b163-4502-98cf-c291899fc22c req-cedf4fed-1609-4049-86aa-3762492472d5 service nova] Acquired lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1280.410211] env[62753]: DEBUG nova.network.neutron [req-fd958215-b163-4502-98cf-c291899fc22c req-cedf4fed-1609-4049-86aa-3762492472d5 service nova] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Refreshing network info cache for port 33f767b7-a237-433b-9476-7cb32f4af66a {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1281.111023] env[62753]: DEBUG nova.network.neutron [req-fd958215-b163-4502-98cf-c291899fc22c req-cedf4fed-1609-4049-86aa-3762492472d5 service nova] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updated VIF entry in instance network info cache for port 33f767b7-a237-433b-9476-7cb32f4af66a. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1281.111412] env[62753]: DEBUG nova.network.neutron [req-fd958215-b163-4502-98cf-c291899fc22c req-cedf4fed-1609-4049-86aa-3762492472d5 service nova] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance_info_cache with network_info: [{"id": "33f767b7-a237-433b-9476-7cb32f4af66a", "address": "fa:16:3e:d0:d1:9b", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33f767b7-a2", "ovs_interfaceid": "33f767b7-a237-433b-9476-7cb32f4af66a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1281.614324] env[62753]: DEBUG oslo_concurrency.lockutils [req-fd958215-b163-4502-98cf-c291899fc22c req-cedf4fed-1609-4049-86aa-3762492472d5 service nova] Releasing lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1308.009806] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1308.010194] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62753) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1309.010898] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1309.011297] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1311.010998] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1311.011395] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1312.006108] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1312.009709] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1312.512676] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1312.513162] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1312.513162] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1312.513376] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62753) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1312.514180] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-88a5f734-b458-4ab6-933c-f23d90901037 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.522713] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e633546f-c6cc-4c81-a2b4-0599171fc43a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.536765] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0983df9e-965b-4dc3-9613-abb903513f4a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.542640] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c4e270a0-0a2e-4c0e-a146-9fbc19a1ed6d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1312.570266] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181426MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=62753) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1312.570397] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1312.570582] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1313.595075] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance 5b196309-161a-4770-90c8-81b57909cb12 actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1313.595318] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1313.595456] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1313.620241] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-225b6af5-ea29-4f89-9c8a-22f5981a3622 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.627810] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-63eb3479-75b3-4ae6-8ace-6caed26e94fe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.657703] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-29a5e3e7-b6a0-4d96-b412-0be0617f6d8a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.664762] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-622d33c9-f841-4231-8f6b-7b9823b87eb3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1313.677580] env[62753]: DEBUG nova.compute.provider_tree [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1314.180919] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1314.685738] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1314.686200] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.115s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1318.540280] env[62753]: DEBUG nova.compute.manager [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Stashing vm_state: active {{(pid=62753) _prep_resize /opt/stack/nova/nova/compute/manager.py:5624}} [ 1318.687032] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1318.687200] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Starting heal instance info cache {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1318.687324] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Rebuilding the list of instances to heal {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1319.061322] env[62753]: DEBUG oslo_concurrency.lockutils [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1319.061587] env[62753]: DEBUG oslo_concurrency.lockutils [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1319.190995] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1319.191182] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquired lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1319.191290] env[62753]: DEBUG nova.network.neutron [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Forcefully refreshing network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1319.191437] env[62753]: DEBUG nova.objects.instance [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lazy-loading 'info_cache' on Instance uuid 5b196309-161a-4770-90c8-81b57909cb12 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1319.566946] env[62753]: INFO nova.compute.claims [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1320.072831] env[62753]: INFO nova.compute.resource_tracker [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating resource usage from migration a463cb1d-d99b-42cd-aee3-b0073de3449f [ 1320.108645] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-23ec2a8c-44b4-4043-9f56-ebb274f19ecd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.115806] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e7c4058-5a64-4710-aabe-12084e53a068 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.144598] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-33e13718-4d82-45f9-8857-87c14d214f80 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.151435] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3a623dc7-8c2a-4aea-9139-85e0f6a1aa26 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1320.163741] env[62753]: DEBUG nova.compute.provider_tree [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1320.666754] env[62753]: DEBUG nova.scheduler.client.report [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1320.909735] env[62753]: DEBUG nova.network.neutron [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance_info_cache with network_info: [{"id": "33f767b7-a237-433b-9476-7cb32f4af66a", "address": "fa:16:3e:d0:d1:9b", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33f767b7-a2", "ovs_interfaceid": "33f767b7-a237-433b-9476-7cb32f4af66a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1321.171775] env[62753]: DEBUG oslo_concurrency.lockutils [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.resize_claim" :: held 2.110s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1321.171954] env[62753]: INFO nova.compute.manager [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Migrating [ 1321.413323] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Releasing lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1321.413520] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updated the network info_cache for instance {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1321.413718] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1321.687818] env[62753]: DEBUG oslo_concurrency.lockutils [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1321.688108] env[62753]: DEBUG oslo_concurrency.lockutils [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1321.688240] env[62753]: DEBUG nova.network.neutron [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1322.381831] env[62753]: DEBUG nova.network.neutron [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance_info_cache with network_info: [{"id": "33f767b7-a237-433b-9476-7cb32f4af66a", "address": "fa:16:3e:d0:d1:9b", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33f767b7-a2", "ovs_interfaceid": "33f767b7-a237-433b-9476-7cb32f4af66a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1322.884873] env[62753]: DEBUG oslo_concurrency.lockutils [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1324.399936] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-9ff1468e-1d2d-41fd-be51-5c469947aa34 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.418016] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance '5b196309-161a-4770-90c8-81b57909cb12' progress to 0 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1324.923867] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1324.924194] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b53d6ca1-0120-4c93-8ae9-7a87d67ffd1a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1324.931912] env[62753]: DEBUG oslo_vmware.api [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1324.931912] env[62753]: value = "task-1333020" [ 1324.931912] env[62753]: _type = "Task" [ 1324.931912] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1324.939533] env[62753]: DEBUG oslo_vmware.api [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333020, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1325.441556] env[62753]: DEBUG oslo_vmware.api [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333020, 'name': PowerOffVM_Task, 'duration_secs': 0.162461} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1325.442051] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1325.442051] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance '5b196309-161a-4770-90c8-81b57909cb12' progress to 17 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1325.948398] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:58Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='84',id=12,is_public=True,memory_mb=256,name='m1.micro',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1325.948670] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1325.948852] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1325.949053] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1325.949212] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1325.949367] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1325.949577] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1325.949744] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1325.949916] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1325.950097] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1325.950278] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1325.955223] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-66059bd3-0a57-43f1-be6d-3378a7e6f423 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1325.970611] env[62753]: DEBUG oslo_vmware.api [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1325.970611] env[62753]: value = "task-1333021" [ 1325.970611] env[62753]: _type = "Task" [ 1325.970611] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1325.979451] env[62753]: DEBUG oslo_vmware.api [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333021, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.481045] env[62753]: DEBUG oslo_vmware.api [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333021, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1326.981354] env[62753]: DEBUG oslo_vmware.api [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333021, 'name': ReconfigVM_Task} progress is 99%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1327.482339] env[62753]: DEBUG oslo_vmware.api [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333021, 'name': ReconfigVM_Task, 'duration_secs': 1.162669} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1327.482739] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance '5b196309-161a-4770-90c8-81b57909cb12' progress to 33 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1327.989747] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1327.990124] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1327.990374] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1327.990634] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1327.990849] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1327.991084] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1327.991346] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1327.991590] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1327.991853] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1327.992088] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1327.992337] env[62753]: DEBUG nova.virt.hardware [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1327.997652] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Reconfiguring VM instance instance-00000071 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1327.998027] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-43ab6858-6991-479e-b1cc-b104b043ffc4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.016171] env[62753]: DEBUG oslo_vmware.api [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1328.016171] env[62753]: value = "task-1333022" [ 1328.016171] env[62753]: _type = "Task" [ 1328.016171] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.023750] env[62753]: DEBUG oslo_vmware.api [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333022, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1328.526017] env[62753]: DEBUG oslo_vmware.api [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333022, 'name': ReconfigVM_Task, 'duration_secs': 0.142653} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1328.526483] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Reconfigured VM instance instance-00000071 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1328.527224] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a129ad4a-a0c2-43c4-84fc-80973176b0c7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.548237] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] 5b196309-161a-4770-90c8-81b57909cb12/5b196309-161a-4770-90c8-81b57909cb12.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1328.548489] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-001cd372-6dca-4b97-9def-5c95f9e90e78 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1328.565108] env[62753]: DEBUG oslo_vmware.api [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1328.565108] env[62753]: value = "task-1333023" [ 1328.565108] env[62753]: _type = "Task" [ 1328.565108] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1328.572214] env[62753]: DEBUG oslo_vmware.api [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333023, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1329.075197] env[62753]: DEBUG oslo_vmware.api [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333023, 'name': ReconfigVM_Task, 'duration_secs': 0.248564} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1329.075554] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Reconfigured VM instance instance-00000071 to attach disk [datastore1] 5b196309-161a-4770-90c8-81b57909cb12/5b196309-161a-4770-90c8-81b57909cb12.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1329.075844] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance '5b196309-161a-4770-90c8-81b57909cb12' progress to 50 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1329.583114] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f9835ffa-2465-43a5-95db-8b21c71b7b6e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.601920] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-79d3686d-67c1-495a-aed3-9a6c8c22f0d6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1329.619558] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance '5b196309-161a-4770-90c8-81b57909cb12' progress to 67 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1330.159892] env[62753]: DEBUG nova.network.neutron [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Port 33f767b7-a237-433b-9476-7cb32f4af66a binding to destination host cpu-1 is already ACTIVE {{(pid=62753) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1331.182308] env[62753]: DEBUG oslo_concurrency.lockutils [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "5b196309-161a-4770-90c8-81b57909cb12-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1331.182670] env[62753]: DEBUG oslo_concurrency.lockutils [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "5b196309-161a-4770-90c8-81b57909cb12-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1331.182670] env[62753]: DEBUG oslo_concurrency.lockutils [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "5b196309-161a-4770-90c8-81b57909cb12-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1332.216716] env[62753]: DEBUG oslo_concurrency.lockutils [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1332.216956] env[62753]: DEBUG oslo_concurrency.lockutils [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1332.217116] env[62753]: DEBUG nova.network.neutron [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1332.915057] env[62753]: DEBUG nova.network.neutron [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance_info_cache with network_info: [{"id": "33f767b7-a237-433b-9476-7cb32f4af66a", "address": "fa:16:3e:d0:d1:9b", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33f767b7-a2", "ovs_interfaceid": "33f767b7-a237-433b-9476-7cb32f4af66a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1333.418701] env[62753]: DEBUG oslo_concurrency.lockutils [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1333.943616] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-5c395b7d-b75d-4126-baf2-8251a3fc59ab {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.963898] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-69d717b0-7fcf-447e-8771-efdc5cd3c681 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1333.970575] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance '5b196309-161a-4770-90c8-81b57909cb12' progress to 83 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1334.476996] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1334.477369] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-4178f92f-56a2-4ac8-bd5b-4a5caecdd233 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1334.484477] env[62753]: DEBUG oslo_vmware.api [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1334.484477] env[62753]: value = "task-1333024" [ 1334.484477] env[62753]: _type = "Task" [ 1334.484477] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1334.491984] env[62753]: DEBUG oslo_vmware.api [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333024, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1334.994138] env[62753]: DEBUG oslo_vmware.api [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333024, 'name': PowerOnVM_Task, 'duration_secs': 0.338923} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1334.994420] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1334.994609] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-741ee5bb-b2f5-4f9a-a29a-4aab970d53ea tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance '5b196309-161a-4770-90c8-81b57909cb12' progress to 100 {{(pid=62753) _update_instance_progress /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1352}} [ 1337.068900] env[62753]: DEBUG nova.network.neutron [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Port 33f767b7-a237-433b-9476-7cb32f4af66a binding to destination host cpu-1 is already ACTIVE {{(pid=62753) migrate_instance_start /opt/stack/nova/nova/network/neutron.py:3228}} [ 1337.069242] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1337.069384] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1337.069553] env[62753]: DEBUG nova.network.neutron [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1337.778799] env[62753]: DEBUG nova.network.neutron [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance_info_cache with network_info: [{"id": "33f767b7-a237-433b-9476-7cb32f4af66a", "address": "fa:16:3e:d0:d1:9b", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33f767b7-a2", "ovs_interfaceid": "33f767b7-a237-433b-9476-7cb32f4af66a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1338.282088] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1338.786026] env[62753]: DEBUG nova.compute.manager [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Hypervisor driver does not support instance shared storage check, assuming it's not on shared storage {{(pid=62753) _is_instance_storage_shared /opt/stack/nova/nova/compute/manager.py:897}} [ 1338.786290] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1338.786521] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1339.290247] env[62753]: DEBUG nova.objects.instance [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lazy-loading 'migration_context' on Instance uuid 5b196309-161a-4770-90c8-81b57909cb12 {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1339.830671] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-13d05085-8393-4191-a5c0-bfd1db8cff08 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.838175] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-55331a96-211f-4133-9f10-c79b51c1dc18 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.866769] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f75d6e3a-062b-46eb-ad1e-1712bc7849d8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.873542] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a79cdac9-dea4-4a62-ad94-836a5a302690 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1339.886422] env[62753]: DEBUG nova.compute.provider_tree [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1340.389947] env[62753]: DEBUG nova.scheduler.client.report [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1341.404215] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.drop_move_claim_at_dest" :: held 2.617s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1342.944534] env[62753]: INFO nova.compute.manager [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Swapping old allocation on dict_keys(['42c8603e-76bc-4e29-ad77-cac0094f7308']) held by migration a463cb1d-d99b-42cd-aee3-b0073de3449f for instance [ 1342.965532] env[62753]: DEBUG nova.scheduler.client.report [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Overwriting current allocation {'allocations': {'42c8603e-76bc-4e29-ad77-cac0094f7308': {'resources': {'DISK_GB': 1, 'MEMORY_MB': 256, 'VCPU': 1}, 'generation': 145}}, 'project_id': '891d1efbcac34ceba5fa72ea6b5fe38b', 'user_id': 'd65bf0df728640ef9104e14d6320ed01', 'consumer_generation': 1} on consumer 5b196309-161a-4770-90c8-81b57909cb12 {{(pid=62753) move_allocations /opt/stack/nova/nova/scheduler/client/report.py:2032}} [ 1343.043204] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1343.043484] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1343.043672] env[62753]: DEBUG nova.network.neutron [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1343.745444] env[62753]: DEBUG nova.network.neutron [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance_info_cache with network_info: [{"id": "33f767b7-a237-433b-9476-7cb32f4af66a", "address": "fa:16:3e:d0:d1:9b", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.8", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap33f767b7-a2", "ovs_interfaceid": "33f767b7-a237-433b-9476-7cb32f4af66a", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1344.248608] env[62753]: DEBUG oslo_concurrency.lockutils [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "refresh_cache-5b196309-161a-4770-90c8-81b57909cb12" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1344.249183] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1344.249478] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-9a6f41bb-8a83-4187-916a-b1a9cb0fe4ab {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.256893] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1344.256893] env[62753]: value = "task-1333025" [ 1344.256893] env[62753]: _type = "Task" [ 1344.256893] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.265327] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333025, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1344.766263] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333025, 'name': PowerOffVM_Task, 'duration_secs': 0.198266} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1344.766536] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1344.767179] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1344.767406] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1344.767568] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1344.767756] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1344.767933] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1344.768109] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1344.768323] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1344.768490] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1344.768660] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1344.768832] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1344.769017] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1344.773923] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-733422be-c20a-403a-b521-85da762bfbb9 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1344.788820] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1344.788820] env[62753]: value = "task-1333026" [ 1344.788820] env[62753]: _type = "Task" [ 1344.788820] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1344.795757] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333026, 'name': ReconfigVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.302695] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333026, 'name': ReconfigVM_Task, 'duration_secs': 0.130522} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1345.304016] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-75980bb8-653d-4649-9790-a3bfeb05e50c {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.332472] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum=,container_format='bare',created_at=,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=1,min_ram=0,name=,owner=,properties=ImageMetaProps,protected=,size=,status=,tags=,updated_at=,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1345.332472] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1345.332859] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1345.332859] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1345.332859] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1345.333037] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1345.333227] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1345.333395] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1345.333578] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1345.333742] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1345.333920] env[62753]: DEBUG nova.virt.hardware [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1345.334718] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-7b117682-ef00-46d8-9d7c-8ec42714e762 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.340227] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1345.340227] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e96ffd-5002-5ede-5245-43575562d352" [ 1345.340227] env[62753]: _type = "Task" [ 1345.340227] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1345.347824] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e96ffd-5002-5ede-5245-43575562d352, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1345.849933] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e96ffd-5002-5ede-5245-43575562d352, 'name': SearchDatastore_Task, 'duration_secs': 0.006869} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1345.855143] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Reconfiguring VM instance instance-00000071 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:122}} [ 1345.855419] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-90b8d1a0-d5aa-4473-b910-878522fe4161 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1345.872698] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1345.872698] env[62753]: value = "task-1333027" [ 1345.872698] env[62753]: _type = "Task" [ 1345.872698] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1345.880208] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333027, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.382618] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333027, 'name': ReconfigVM_Task, 'duration_secs': 0.178472} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.383028] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Reconfigured VM instance instance-00000071 to detach disk 2000 {{(pid=62753) detach_disk_from_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:127}} [ 1346.383719] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1b8a93ae-8300-4fdc-95b5-125a3b2c43c2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.404916] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Reconfiguring VM instance instance-00000071 to attach disk [datastore1] 5b196309-161a-4770-90c8-81b57909cb12/5b196309-161a-4770-90c8-81b57909cb12.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1346.405179] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-d4770abd-e318-4906-a773-213214e762e2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.422475] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1346.422475] env[62753]: value = "task-1333028" [ 1346.422475] env[62753]: _type = "Task" [ 1346.422475] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1346.429709] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333028, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1346.933241] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333028, 'name': ReconfigVM_Task, 'duration_secs': 0.252221} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1346.933543] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Reconfigured VM instance instance-00000071 to attach disk [datastore1] 5b196309-161a-4770-90c8-81b57909cb12/5b196309-161a-4770-90c8-81b57909cb12.vmdk or device None with type thin {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1346.934369] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-6a8363d6-2dba-4785-b7a0-151b4876c650 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.951777] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-842593bf-008d-42f5-8163-8c9a879d2a60 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.968853] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-998a89cc-d4a6-4b63-be0d-0e7e84a68e5a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.987499] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a3fea1b7-7e96-42b8-b217-00c3e3229c42 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.993850] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1346.994115] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-66d680a1-e37d-488a-8a32-3fba843eef93 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1346.999250] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1346.999250] env[62753]: value = "task-1333029" [ 1346.999250] env[62753]: _type = "Task" [ 1346.999250] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1347.006062] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333029, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1347.509644] env[62753]: DEBUG oslo_vmware.api [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333029, 'name': PowerOnVM_Task, 'duration_secs': 0.349028} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1347.510062] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1348.520361] env[62753]: INFO nova.compute.manager [None req-5c07b0d6-f786-458c-aed6-e12d3515d944 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance to original state: 'active' [ 1349.442665] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "5b196309-161a-4770-90c8-81b57909cb12" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1349.443044] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "5b196309-161a-4770-90c8-81b57909cb12" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1349.443183] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "5b196309-161a-4770-90c8-81b57909cb12-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1349.443382] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "5b196309-161a-4770-90c8-81b57909cb12-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1349.443565] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "5b196309-161a-4770-90c8-81b57909cb12-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1349.445815] env[62753]: INFO nova.compute.manager [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Terminating instance [ 1349.447611] env[62753]: DEBUG nova.compute.manager [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1349.447810] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1349.448697] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8d6d12ba-5883-427a-a682-434e272185ec {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.456385] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1349.456899] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-08a8b58a-355c-461b-a571-167d278c529e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1349.462654] env[62753]: DEBUG oslo_vmware.api [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1349.462654] env[62753]: value = "task-1333030" [ 1349.462654] env[62753]: _type = "Task" [ 1349.462654] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1349.470174] env[62753]: DEBUG oslo_vmware.api [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333030, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1349.972856] env[62753]: DEBUG oslo_vmware.api [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333030, 'name': PowerOffVM_Task, 'duration_secs': 0.179414} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1349.973208] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1349.973320] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1349.973572] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-a664ed9c-574c-4520-a758-864623ff0ac4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.035604] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1350.035942] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Deleting contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1350.036194] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Deleting the datastore file [datastore1] 5b196309-161a-4770-90c8-81b57909cb12 {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1350.036493] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-c0a9b0ce-d9b0-4ec7-97f2-ae1363f8b9b7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1350.042067] env[62753]: DEBUG oslo_vmware.api [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1350.042067] env[62753]: value = "task-1333032" [ 1350.042067] env[62753]: _type = "Task" [ 1350.042067] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1350.049696] env[62753]: DEBUG oslo_vmware.api [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333032, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1350.552530] env[62753]: DEBUG oslo_vmware.api [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333032, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.145945} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1350.552775] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1350.552917] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Deleted contents of the VM from datastore datastore1 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1350.553114] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1350.553297] env[62753]: INFO nova.compute.manager [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Took 1.11 seconds to destroy the instance on the hypervisor. [ 1350.553545] env[62753]: DEBUG oslo.service.loopingcall [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1350.553735] env[62753]: DEBUG nova.compute.manager [-] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1350.553831] env[62753]: DEBUG nova.network.neutron [-] [instance: 5b196309-161a-4770-90c8-81b57909cb12] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1350.994631] env[62753]: DEBUG nova.compute.manager [req-4517c736-4cca-477b-b15a-8ca1b9be8956 req-be2497af-fe3e-47c0-ab0b-9926f4034318 service nova] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Received event network-vif-deleted-33f767b7-a237-433b-9476-7cb32f4af66a {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1350.994983] env[62753]: INFO nova.compute.manager [req-4517c736-4cca-477b-b15a-8ca1b9be8956 req-be2497af-fe3e-47c0-ab0b-9926f4034318 service nova] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Neutron deleted interface 33f767b7-a237-433b-9476-7cb32f4af66a; detaching it from the instance and deleting it from the info cache [ 1350.995393] env[62753]: DEBUG nova.network.neutron [req-4517c736-4cca-477b-b15a-8ca1b9be8956 req-be2497af-fe3e-47c0-ab0b-9926f4034318 service nova] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1351.466466] env[62753]: DEBUG nova.network.neutron [-] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1351.499072] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-6257f900-d4ff-4db8-8247-0736a14138fe {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.508821] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a69aced6-4118-40b1-b9e8-6a0ebbf8f26e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1351.529748] env[62753]: DEBUG nova.compute.manager [req-4517c736-4cca-477b-b15a-8ca1b9be8956 req-be2497af-fe3e-47c0-ab0b-9926f4034318 service nova] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Detach interface failed, port_id=33f767b7-a237-433b-9476-7cb32f4af66a, reason: Instance 5b196309-161a-4770-90c8-81b57909cb12 could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1351.969087] env[62753]: INFO nova.compute.manager [-] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Took 1.41 seconds to deallocate network for instance. [ 1352.476507] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1352.476825] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1352.476928] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1352.495911] env[62753]: INFO nova.scheduler.client.report [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Deleted allocations for instance 5b196309-161a-4770-90c8-81b57909cb12 [ 1353.003941] env[62753]: DEBUG oslo_concurrency.lockutils [None req-8f4801ce-e3b1-48c9-9d3d-76711925b992 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "5b196309-161a-4770-90c8-81b57909cb12" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 3.561s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1354.443728] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1354.444028] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" acquired by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1354.946055] env[62753]: DEBUG nova.compute.manager [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Starting instance... {{(pid=62753) _do_build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2432}} [ 1355.467043] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1355.467318] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1355.468854] env[62753]: INFO nova.compute.claims [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Claim successful on node domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 [ 1356.506512] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ada5b8df-ca7f-4eba-965f-c2f98d10fb42 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.514518] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2501e8a5-9871-40d0-99d5-94453e1f2860 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.544129] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c159a5c5-6011-44bf-990a-c44183420b68 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.551653] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-e4ef5ab8-2a54-4992-a8a0-1d8cfa56329e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1356.564741] env[62753]: DEBUG nova.compute.provider_tree [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1357.068091] env[62753]: DEBUG nova.scheduler.client.report [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1357.574177] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.instance_claim" :: held 2.106s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1357.574975] env[62753]: DEBUG nova.compute.manager [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Start building networks asynchronously for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2829}} [ 1358.081614] env[62753]: DEBUG nova.compute.utils [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Using /dev/sd instead of None {{(pid=62753) get_next_device_name /opt/stack/nova/nova/compute/utils.py:238}} [ 1358.083066] env[62753]: DEBUG nova.compute.manager [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Allocating IP information in the background. {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1981}} [ 1358.083208] env[62753]: DEBUG nova.network.neutron [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] allocate_for_instance() {{(pid=62753) allocate_for_instance /opt/stack/nova/nova/network/neutron.py:1205}} [ 1358.131739] env[62753]: DEBUG nova.policy [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Policy check for network:attach_external_network failed with credentials {'is_admin': False, 'user_id': 'd65bf0df728640ef9104e14d6320ed01', 'user_domain_id': 'default', 'system_scope': None, 'domain_id': None, 'project_id': '891d1efbcac34ceba5fa72ea6b5fe38b', 'project_domain_id': 'default', 'roles': ['member', 'reader'], 'is_admin_project': True, 'service_user_id': None, 'service_user_domain_id': None, 'service_project_id': None, 'service_project_domain_id': None, 'service_roles': []} {{(pid=62753) authorize /opt/stack/nova/nova/policy.py:201}} [ 1358.396326] env[62753]: DEBUG nova.network.neutron [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Successfully created port: 176523fa-3534-4c71-a62e-d8ce87b429de {{(pid=62753) _create_port_minimal /opt/stack/nova/nova/network/neutron.py:548}} [ 1358.586272] env[62753]: DEBUG nova.compute.manager [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Start building block device mappings for instance. {{(pid=62753) _build_resources /opt/stack/nova/nova/compute/manager.py:2864}} [ 1359.596346] env[62753]: DEBUG nova.compute.manager [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Start spawning the instance on the hypervisor. {{(pid=62753) _build_and_run_instance /opt/stack/nova/nova/compute/manager.py:2638}} [ 1359.620746] env[62753]: DEBUG nova.virt.hardware [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Getting desirable topologies for flavor Flavor(created_at=2024-10-21T20:24:55Z,deleted=False,deleted_at=None,description=None,disabled=False,ephemeral_gb=0,extra_specs={hw_rng:allowed='True'},flavorid='42',id=11,is_public=True,memory_mb=192,name='m1.nano',projects=,root_gb=1,rxtx_factor=1.0,swap=0,updated_at=None,vcpu_weight=0,vcpus=1) and image_meta ImageMeta(checksum='879b8eacf4b511d84bac79c7fe0e0d0a',container_format='bare',created_at=2024-10-21T20:24:38Z,direct_url=,disk_format='vmdk',id=a7541f15-bad6-4ea8-95ce-3499a4f01dda,min_disk=0,min_ram=0,name='cirros-d240228-sparse;paraVirtual;vmxnet3',owner='73cbe287fdab4fb3b03f5229409899f4',properties=ImageMetaProps,protected=,size=21318656,status='active',tags=,updated_at=2024-10-21T20:24:39Z,virtual_size=,visibility=), allow threads: False {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:563}} [ 1359.621015] env[62753]: DEBUG nova.virt.hardware [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:348}} [ 1359.621193] env[62753]: DEBUG nova.virt.hardware [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image limits 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:352}} [ 1359.621383] env[62753]: DEBUG nova.virt.hardware [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Flavor pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:388}} [ 1359.621534] env[62753]: DEBUG nova.virt.hardware [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Image pref 0:0:0 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:392}} [ 1359.621688] env[62753]: DEBUG nova.virt.hardware [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Chose sockets=0, cores=0, threads=0; limits were sockets=65536, cores=65536, threads=65536 {{(pid=62753) get_cpu_topology_constraints /opt/stack/nova/nova/virt/hardware.py:430}} [ 1359.621898] env[62753]: DEBUG nova.virt.hardware [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Topology preferred VirtCPUTopology(cores=0,sockets=0,threads=0), maximum VirtCPUTopology(cores=65536,sockets=65536,threads=65536) {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:569}} [ 1359.622077] env[62753]: DEBUG nova.virt.hardware [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Build topologies for 1 vcpu(s) 1:1:1 {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:471}} [ 1359.622254] env[62753]: DEBUG nova.virt.hardware [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Got 1 possible topologies {{(pid=62753) _get_possible_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:501}} [ 1359.622419] env[62753]: DEBUG nova.virt.hardware [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Possible topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:575}} [ 1359.622595] env[62753]: DEBUG nova.virt.hardware [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Sorted desired topologies [VirtCPUTopology(cores=1,sockets=1,threads=1)] {{(pid=62753) _get_desirable_cpu_topologies /opt/stack/nova/nova/virt/hardware.py:577}} [ 1359.623473] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f40c5012-84b8-4f0d-924f-0a37e5bc7688 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.631373] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8b6c36ca-e9cf-4b1c-915f-1866f0c8a9b2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1359.935163] env[62753]: DEBUG nova.compute.manager [req-0ca8f0e6-88b6-4797-af3c-d10b0327e033 req-8a0775c3-ca9c-47b1-a99b-f6b7cb9a1a64 service nova] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Received event network-vif-plugged-176523fa-3534-4c71-a62e-d8ce87b429de {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1359.935393] env[62753]: DEBUG oslo_concurrency.lockutils [req-0ca8f0e6-88b6-4797-af3c-d10b0327e033 req-8a0775c3-ca9c-47b1-a99b-f6b7cb9a1a64 service nova] Acquiring lock "a3053e3f-9dc7-4724-bd6d-4deaf4ee931c-events" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1359.935639] env[62753]: DEBUG oslo_concurrency.lockutils [req-0ca8f0e6-88b6-4797-af3c-d10b0327e033 req-8a0775c3-ca9c-47b1-a99b-f6b7cb9a1a64 service nova] Lock "a3053e3f-9dc7-4724-bd6d-4deaf4ee931c-events" acquired by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1359.935808] env[62753]: DEBUG oslo_concurrency.lockutils [req-0ca8f0e6-88b6-4797-af3c-d10b0327e033 req-8a0775c3-ca9c-47b1-a99b-f6b7cb9a1a64 service nova] Lock "a3053e3f-9dc7-4724-bd6d-4deaf4ee931c-events" "released" by "nova.compute.manager.InstanceEvents.pop_instance_event.._pop_event" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1359.935926] env[62753]: DEBUG nova.compute.manager [req-0ca8f0e6-88b6-4797-af3c-d10b0327e033 req-8a0775c3-ca9c-47b1-a99b-f6b7cb9a1a64 service nova] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] No waiting events found dispatching network-vif-plugged-176523fa-3534-4c71-a62e-d8ce87b429de {{(pid=62753) pop_instance_event /opt/stack/nova/nova/compute/manager.py:320}} [ 1359.936112] env[62753]: WARNING nova.compute.manager [req-0ca8f0e6-88b6-4797-af3c-d10b0327e033 req-8a0775c3-ca9c-47b1-a99b-f6b7cb9a1a64 service nova] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Received unexpected event network-vif-plugged-176523fa-3534-4c71-a62e-d8ce87b429de for instance with vm_state building and task_state spawning. [ 1360.015244] env[62753]: DEBUG nova.network.neutron [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Successfully updated port: 176523fa-3534-4c71-a62e-d8ce87b429de {{(pid=62753) _update_port /opt/stack/nova/nova/network/neutron.py:586}} [ 1360.517971] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1360.518218] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1360.518407] env[62753]: DEBUG nova.network.neutron [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1361.049794] env[62753]: DEBUG nova.network.neutron [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Instance cache missing network info. {{(pid=62753) _get_preexisting_port_ids /opt/stack/nova/nova/network/neutron.py:3380}} [ 1361.170691] env[62753]: DEBUG nova.network.neutron [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Updating instance_info_cache with network_info: [{"id": "176523fa-3534-4c71-a62e-d8ce87b429de", "address": "fa:16:3e:90:9d:46", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap176523fa-35", "ovs_interfaceid": "176523fa-3534-4c71-a62e-d8ce87b429de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1361.672964] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1361.673279] env[62753]: DEBUG nova.compute.manager [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Instance network_info: |[{"id": "176523fa-3534-4c71-a62e-d8ce87b429de", "address": "fa:16:3e:90:9d:46", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap176523fa-35", "ovs_interfaceid": "176523fa-3534-4c71-a62e-d8ce87b429de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}]| {{(pid=62753) _allocate_network_async /opt/stack/nova/nova/compute/manager.py:1996}} [ 1361.673754] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Instance VIF info [{'network_name': 'br-int', 'mac_address': 'fa:16:3e:90:9d:46', 'network_ref': {'type': 'OpaqueNetwork', 'network-id': 'cae1d6a8-cbba-4bbf-af10-ba5467340475', 'network-type': 'nsx.LogicalSwitch', 'use-external-id': True}, 'iface_id': '176523fa-3534-4c71-a62e-d8ce87b429de', 'vif_model': 'vmxnet3'}] {{(pid=62753) build_virtual_machine /opt/stack/nova/nova/virt/vmwareapi/vmops.py:279}} [ 1361.681214] env[62753]: DEBUG oslo.service.loopingcall [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for function nova.virt.vmwareapi.vm_util.create_vm to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1361.681423] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Creating VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1337}} [ 1361.681652] env[62753]: DEBUG oslo_vmware.service [-] Invoking Folder.CreateVM_Task with opID=oslo.vmware-f6835584-51e9-4742-9d38-d1da2f093f98 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1361.701709] env[62753]: DEBUG oslo_vmware.api [-] Waiting for the task: (returnval){ [ 1361.701709] env[62753]: value = "task-1333033" [ 1361.701709] env[62753]: _type = "Task" [ 1361.701709] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1361.709042] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333033, 'name': CreateVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1361.960859] env[62753]: DEBUG nova.compute.manager [req-4d4605f7-fd92-4f97-aa9c-07500f7857d2 req-d37be26f-8d9f-4917-8f06-77cd85555a3c service nova] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Received event network-changed-176523fa-3534-4c71-a62e-d8ce87b429de {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1361.960859] env[62753]: DEBUG nova.compute.manager [req-4d4605f7-fd92-4f97-aa9c-07500f7857d2 req-d37be26f-8d9f-4917-8f06-77cd85555a3c service nova] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Refreshing instance network info cache due to event network-changed-176523fa-3534-4c71-a62e-d8ce87b429de. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1361.961115] env[62753]: DEBUG oslo_concurrency.lockutils [req-4d4605f7-fd92-4f97-aa9c-07500f7857d2 req-d37be26f-8d9f-4917-8f06-77cd85555a3c service nova] Acquiring lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1361.961273] env[62753]: DEBUG oslo_concurrency.lockutils [req-4d4605f7-fd92-4f97-aa9c-07500f7857d2 req-d37be26f-8d9f-4917-8f06-77cd85555a3c service nova] Acquired lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1361.961547] env[62753]: DEBUG nova.network.neutron [req-4d4605f7-fd92-4f97-aa9c-07500f7857d2 req-d37be26f-8d9f-4917-8f06-77cd85555a3c service nova] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Refreshing network info cache for port 176523fa-3534-4c71-a62e-d8ce87b429de {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1362.212414] env[62753]: DEBUG oslo_vmware.api [-] Task: {'id': task-1333033, 'name': CreateVM_Task, 'duration_secs': 0.287618} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1362.212770] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [-] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Created VM on the ESX host {{(pid=62753) create_vm /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1359}} [ 1362.213238] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1362.213414] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1362.213747] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired external semaphore "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:321}} [ 1362.213991] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-c9153306-465d-4dc3-8bad-1cc0a2a7ecb8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.218459] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1362.218459] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e047f8-8e70-7f58-280b-ab276881a0a5" [ 1362.218459] env[62753]: _type = "Task" [ 1362.218459] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1362.227141] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e047f8-8e70-7f58-280b-ab276881a0a5, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1362.640379] env[62753]: DEBUG nova.network.neutron [req-4d4605f7-fd92-4f97-aa9c-07500f7857d2 req-d37be26f-8d9f-4917-8f06-77cd85555a3c service nova] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Updated VIF entry in instance network info cache for port 176523fa-3534-4c71-a62e-d8ce87b429de. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1362.640706] env[62753]: DEBUG nova.network.neutron [req-4d4605f7-fd92-4f97-aa9c-07500f7857d2 req-d37be26f-8d9f-4917-8f06-77cd85555a3c service nova] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Updating instance_info_cache with network_info: [{"id": "176523fa-3534-4c71-a62e-d8ce87b429de", "address": "fa:16:3e:90:9d:46", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": []}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap176523fa-35", "ovs_interfaceid": "176523fa-3534-4c71-a62e-d8ce87b429de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1362.728572] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52e047f8-8e70-7f58-280b-ab276881a0a5, 'name': SearchDatastore_Task, 'duration_secs': 0.012404} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1362.728865] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1362.729152] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Processing image a7541f15-bad6-4ea8-95ce-3499a4f01dda {{(pid=62753) _fetch_image_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:624}} [ 1362.729384] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1362.729537] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1362.729720] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Creating directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:399}} [ 1362.730011] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.MakeDirectory with opID=oslo.vmware-2ac555c3-5b04-4ae8-a23d-8218a51e62f1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.737075] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Created directory with path [datastore2] devstack-image-cache_base {{(pid=62753) mkdir /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:404}} [ 1362.737252] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Folder [datastore2] devstack-image-cache_base created. {{(pid=62753) _create_folder_if_missing /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1845}} [ 1362.737930] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31bf6a76-f823-47a1-add9-ae028ae43347 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1362.742349] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1362.742349] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bebdab-7382-dcc7-e9a5-663cd94b262d" [ 1362.742349] env[62753]: _type = "Task" [ 1362.742349] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1362.749169] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bebdab-7382-dcc7-e9a5-663cd94b262d, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1363.143647] env[62753]: DEBUG oslo_concurrency.lockutils [req-4d4605f7-fd92-4f97-aa9c-07500f7857d2 req-d37be26f-8d9f-4917-8f06-77cd85555a3c service nova] Releasing lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1363.252912] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52bebdab-7382-dcc7-e9a5-663cd94b262d, 'name': SearchDatastore_Task, 'duration_secs': 0.014138} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1363.253691] env[62753]: DEBUG oslo_vmware.service [-] Invoking HostDatastoreBrowser.SearchDatastore_Task with opID=oslo.vmware-31951c70-bbb9-42e0-a59c-e56b10c30fbd {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.258614] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1363.258614] env[62753]: value = "session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52763641-4a6d-bac9-59fc-8603ddaffa36" [ 1363.258614] env[62753]: _type = "Task" [ 1363.258614] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1363.265515] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52763641-4a6d-bac9-59fc-8603ddaffa36, 'name': SearchDatastore_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1363.769168] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': session[52a01f06-114e-090d-20d9-6dfbaa4cd932]52763641-4a6d-bac9-59fc-8603ddaffa36, 'name': SearchDatastore_Task, 'duration_secs': 0.008625} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1363.769406] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "[datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1363.769673] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Copying Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] a3053e3f-9dc7-4724-bd6d-4deaf4ee931c/a3053e3f-9dc7-4724-bd6d-4deaf4ee931c.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1423}} [ 1363.769925] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.CopyVirtualDisk_Task with opID=oslo.vmware-dc917618-00b1-4fcf-ae4c-78b104750950 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1363.776442] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1363.776442] env[62753]: value = "task-1333034" [ 1363.776442] env[62753]: _type = "Task" [ 1363.776442] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1363.783391] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333034, 'name': CopyVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1364.286738] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333034, 'name': CopyVirtualDisk_Task, 'duration_secs': 0.412404} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1364.287134] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Copied Virtual Disk [datastore2] devstack-image-cache_base/a7541f15-bad6-4ea8-95ce-3499a4f01dda/a7541f15-bad6-4ea8-95ce-3499a4f01dda.vmdk to [datastore2] a3053e3f-9dc7-4724-bd6d-4deaf4ee931c/a3053e3f-9dc7-4724-bd6d-4deaf4ee931c.vmdk {{(pid=62753) copy_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1434}} [ 1364.287251] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Extending root virtual disk to 1048576 {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:167}} [ 1364.287503] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualDiskManager.ExtendVirtualDisk_Task with opID=oslo.vmware-d23e4817-10fb-48fd-ac36-265de4aaaadc {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.293330] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1364.293330] env[62753]: value = "task-1333035" [ 1364.293330] env[62753]: _type = "Task" [ 1364.293330] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1364.300735] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333035, 'name': ExtendVirtualDisk_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1364.803499] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333035, 'name': ExtendVirtualDisk_Task, 'duration_secs': 0.055689} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1364.803784] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Extended root virtual disk {{(pid=62753) _extend_virtual_disk /opt/stack/nova/nova/virt/vmwareapi/vmops.py:189}} [ 1364.804632] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a39a3755-69c6-4297-adc7-c13ba8a08959 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.826791] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Reconfiguring VM instance instance-00000072 to attach disk [datastore2] a3053e3f-9dc7-4724-bd6d-4deaf4ee931c/a3053e3f-9dc7-4724-bd6d-4deaf4ee931c.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:81}} [ 1364.827043] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.ReconfigVM_Task with opID=oslo.vmware-810b5c28-07c1-4c94-8cb9-b2fd1c9c5a0d {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1364.845274] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1364.845274] env[62753]: value = "task-1333036" [ 1364.845274] env[62753]: _type = "Task" [ 1364.845274] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1364.853711] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333036, 'name': ReconfigVM_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1365.355310] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333036, 'name': ReconfigVM_Task, 'duration_secs': 0.259254} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1365.355692] env[62753]: DEBUG nova.virt.vmwareapi.volumeops [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Reconfigured VM instance instance-00000072 to attach disk [datastore2] a3053e3f-9dc7-4724-bd6d-4deaf4ee931c/a3053e3f-9dc7-4724-bd6d-4deaf4ee931c.vmdk or device None with type sparse {{(pid=62753) attach_disk_to_vm /opt/stack/nova/nova/virt/vmwareapi/volumeops.py:88}} [ 1365.356226] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.Rename_Task with opID=oslo.vmware-1a2571a4-c425-46f8-9cd8-e833987d6312 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.362449] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1365.362449] env[62753]: value = "task-1333037" [ 1365.362449] env[62753]: _type = "Task" [ 1365.362449] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1365.369873] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333037, 'name': Rename_Task} progress is 5%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1365.872493] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333037, 'name': Rename_Task, 'duration_secs': 0.12573} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1365.872864] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1365.872962] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-bd331694-f148-46db-be37-d413d6b2d0e0 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1365.879786] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1365.879786] env[62753]: value = "task-1333038" [ 1365.879786] env[62753]: _type = "Task" [ 1365.879786] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1365.887288] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333038, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1366.391097] env[62753]: DEBUG oslo_vmware.api [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333038, 'name': PowerOnVM_Task, 'duration_secs': 0.41364} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1366.391476] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1366.391692] env[62753]: INFO nova.compute.manager [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Took 6.80 seconds to spawn the instance on the hypervisor. [ 1366.391924] env[62753]: DEBUG nova.compute.manager [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1366.392796] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7616641b-273a-4158-9f80-bbc62e19b185 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1366.910240] env[62753]: INFO nova.compute.manager [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Took 11.46 seconds to build instance. [ 1367.409448] env[62753]: DEBUG nova.compute.manager [req-95a09d5d-ab6a-4085-9721-3fb1cde2ffae req-a98f49b8-ce6a-4ca3-ba71-54184fa3f984 service nova] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Received event network-changed-176523fa-3534-4c71-a62e-d8ce87b429de {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1367.409685] env[62753]: DEBUG nova.compute.manager [req-95a09d5d-ab6a-4085-9721-3fb1cde2ffae req-a98f49b8-ce6a-4ca3-ba71-54184fa3f984 service nova] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Refreshing instance network info cache due to event network-changed-176523fa-3534-4c71-a62e-d8ce87b429de. {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11136}} [ 1367.409846] env[62753]: DEBUG oslo_concurrency.lockutils [req-95a09d5d-ab6a-4085-9721-3fb1cde2ffae req-a98f49b8-ce6a-4ca3-ba71-54184fa3f984 service nova] Acquiring lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1367.410008] env[62753]: DEBUG oslo_concurrency.lockutils [req-95a09d5d-ab6a-4085-9721-3fb1cde2ffae req-a98f49b8-ce6a-4ca3-ba71-54184fa3f984 service nova] Acquired lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1367.410225] env[62753]: DEBUG nova.network.neutron [req-95a09d5d-ab6a-4085-9721-3fb1cde2ffae req-a98f49b8-ce6a-4ca3-ba71-54184fa3f984 service nova] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Refreshing network info cache for port 176523fa-3534-4c71-a62e-d8ce87b429de {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2064}} [ 1367.412926] env[62753]: DEBUG oslo_concurrency.lockutils [None req-ea4de0b0-d090-4416-bc82-f8dfc9edecab tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" "released" by "nova.compute.manager.ComputeManager.build_and_run_instance.._locked_do_build_and_run_instance" :: held 12.969s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1368.011201] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1368.011376] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62753) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1368.115867] env[62753]: DEBUG nova.network.neutron [req-95a09d5d-ab6a-4085-9721-3fb1cde2ffae req-a98f49b8-ce6a-4ca3-ba71-54184fa3f984 service nova] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Updated VIF entry in instance network info cache for port 176523fa-3534-4c71-a62e-d8ce87b429de. {{(pid=62753) _build_network_info_model /opt/stack/nova/nova/network/neutron.py:3539}} [ 1368.116273] env[62753]: DEBUG nova.network.neutron [req-95a09d5d-ab6a-4085-9721-3fb1cde2ffae req-a98f49b8-ce6a-4ca3-ba71-54184fa3f984 service nova] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Updating instance_info_cache with network_info: [{"id": "176523fa-3534-4c71-a62e-d8ce87b429de", "address": "fa:16:3e:90:9d:46", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap176523fa-35", "ovs_interfaceid": "176523fa-3534-4c71-a62e-d8ce87b429de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1368.619308] env[62753]: DEBUG oslo_concurrency.lockutils [req-95a09d5d-ab6a-4085-9721-3fb1cde2ffae req-a98f49b8-ce6a-4ca3-ba71-54184fa3f984 service nova] Releasing lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1371.011535] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1371.012076] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1371.012076] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1372.010693] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1373.005857] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1373.006276] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._sync_scheduler_instance_info {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1374.010975] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1374.514151] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1374.514405] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1374.514578] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1374.514741] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62753) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1374.515698] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4c095fca-0070-44b7-a674-80e39564ffaa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.524346] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-ec00756b-e3fd-4825-bf3c-23461c2360a6 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.538166] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-7c5431e8-3471-472e-8755-53d3bc16d2fa {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.544325] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-be081ac2-15bf-4c5c-bbc3-f3d3bf819209 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1374.571834] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181429MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=62753) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1374.571974] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1374.572171] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1375.596411] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Instance a3053e3f-9dc7-4724-bd6d-4deaf4ee931c actively managed on this compute host and has allocations in placement: {'resources': {'DISK_GB': 1, 'MEMORY_MB': 192, 'VCPU': 1}}. {{(pid=62753) _remove_deleted_instances_allocations /opt/stack/nova/nova/compute/resource_tracker.py:1707}} [ 1375.596671] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Total usable vcpus: 48, total allocated vcpus: 1 {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1375.596874] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=704MB phys_disk=200GB used_disk=1GB total_vcpus=48 used_vcpus=1 pci_stats=[] {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1375.622916] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-0fbbba1c-9786-402d-b4fa-3c3004e9740e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.630392] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1ebcb270-7aef-40ef-9d4b-fce9e7633a59 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.660536] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-89d39792-a112-421b-9e32-127f14e72861 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.668074] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-aa81d1d0-42af-4f89-8a4b-494d87eb40ca {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1375.681143] env[62753]: DEBUG nova.compute.provider_tree [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1376.184703] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1376.690337] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1376.690727] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.118s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1379.690941] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1379.691406] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Starting heal instance info cache {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1379.691406] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Rebuilding the list of instances to heal {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1380.220656] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1380.220810] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquired lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1380.220961] env[62753]: DEBUG nova.network.neutron [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Forcefully refreshing network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2061}} [ 1380.221139] env[62753]: DEBUG nova.objects.instance [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lazy-loading 'info_cache' on Instance uuid a3053e3f-9dc7-4724-bd6d-4deaf4ee931c {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1381.941649] env[62753]: DEBUG nova.network.neutron [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Updating instance_info_cache with network_info: [{"id": "176523fa-3534-4c71-a62e-d8ce87b429de", "address": "fa:16:3e:90:9d:46", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap176523fa-35", "ovs_interfaceid": "176523fa-3534-4c71-a62e-d8ce87b429de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1382.444282] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Releasing lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1382.444504] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Updated the network info_cache for instance {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9999}} [ 1382.444710] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rebooting_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1404.086479] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2d5d8ca9-3e9e-451d-bb11-c8628f81b83e tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1404.086834] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2d5d8ca9-3e9e-451d-bb11-c8628f81b83e tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" acquired by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1404.086967] env[62753]: DEBUG nova.compute.manager [None req-2d5d8ca9-3e9e-451d-bb11-c8628f81b83e tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1404.087921] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-fb6fbb0e-c9e4-4dfa-9c4f-666a7bcd5755 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.094841] env[62753]: DEBUG nova.compute.manager [None req-2d5d8ca9-3e9e-451d-bb11-c8628f81b83e tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Stopping instance; current vm_state: active, current task_state: powering-off, current DB power_state: 1, current VM power_state: 1 {{(pid=62753) do_stop_instance /opt/stack/nova/nova/compute/manager.py:3368}} [ 1404.095404] env[62753]: DEBUG nova.objects.instance [None req-2d5d8ca9-3e9e-451d-bb11-c8628f81b83e tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lazy-loading 'flavor' on Instance uuid a3053e3f-9dc7-4724-bd6d-4deaf4ee931c {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1404.602074] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d5d8ca9-3e9e-451d-bb11-c8628f81b83e tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1404.602400] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-b2c4a1fa-802d-4fbe-ae87-85ceb1d2fcf2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1404.609817] env[62753]: DEBUG oslo_vmware.api [None req-2d5d8ca9-3e9e-451d-bb11-c8628f81b83e tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1404.609817] env[62753]: value = "task-1333039" [ 1404.609817] env[62753]: _type = "Task" [ 1404.609817] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1404.617821] env[62753]: DEBUG oslo_vmware.api [None req-2d5d8ca9-3e9e-451d-bb11-c8628f81b83e tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333039, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1405.119715] env[62753]: DEBUG oslo_vmware.api [None req-2d5d8ca9-3e9e-451d-bb11-c8628f81b83e tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333039, 'name': PowerOffVM_Task, 'duration_secs': 0.168966} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1405.120103] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-2d5d8ca9-3e9e-451d-bb11-c8628f81b83e tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1405.120298] env[62753]: DEBUG nova.compute.manager [None req-2d5d8ca9-3e9e-451d-bb11-c8628f81b83e tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1405.121084] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-932c8f30-0182-4394-86da-b2adfa40a712 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1405.631975] env[62753]: DEBUG oslo_concurrency.lockutils [None req-2d5d8ca9-3e9e-451d-bb11-c8628f81b83e tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" "released" by "nova.compute.manager.ComputeManager.stop_instance..do_stop_instance" :: held 1.545s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1406.446866] env[62753]: DEBUG nova.objects.instance [None req-62c1e732-39f8-407f-9e8d-e93f184a3c2b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lazy-loading 'flavor' on Instance uuid a3053e3f-9dc7-4724-bd6d-4deaf4ee931c {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1406.952550] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62c1e732-39f8-407f-9e8d-e93f184a3c2b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1406.952745] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62c1e732-39f8-407f-9e8d-e93f184a3c2b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1406.952933] env[62753]: DEBUG nova.network.neutron [None req-62c1e732-39f8-407f-9e8d-e93f184a3c2b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1406.953137] env[62753]: DEBUG nova.objects.instance [None req-62c1e732-39f8-407f-9e8d-e93f184a3c2b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lazy-loading 'info_cache' on Instance uuid a3053e3f-9dc7-4724-bd6d-4deaf4ee931c {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1407.456680] env[62753]: DEBUG nova.objects.base [None req-62c1e732-39f8-407f-9e8d-e93f184a3c2b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Object Instance lazy-loaded attributes: flavor,info_cache {{(pid=62753) wrapper /opt/stack/nova/nova/objects/base.py:126}} [ 1408.153106] env[62753]: DEBUG nova.network.neutron [None req-62c1e732-39f8-407f-9e8d-e93f184a3c2b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Updating instance_info_cache with network_info: [{"id": "176523fa-3534-4c71-a62e-d8ce87b429de", "address": "fa:16:3e:90:9d:46", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap176523fa-35", "ovs_interfaceid": "176523fa-3534-4c71-a62e-d8ce87b429de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1408.656378] env[62753]: DEBUG oslo_concurrency.lockutils [None req-62c1e732-39f8-407f-9e8d-e93f184a3c2b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1409.160405] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-62c1e732-39f8-407f-9e8d-e93f184a3c2b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Powering on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1452}} [ 1409.160687] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-5209e017-4287-4b75-a4c2-b0e79c981b27 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1409.168052] env[62753]: DEBUG oslo_vmware.api [None req-62c1e732-39f8-407f-9e8d-e93f184a3c2b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1409.168052] env[62753]: value = "task-1333040" [ 1409.168052] env[62753]: _type = "Task" [ 1409.168052] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1409.175258] env[62753]: DEBUG oslo_vmware.api [None req-62c1e732-39f8-407f-9e8d-e93f184a3c2b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333040, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1409.677328] env[62753]: DEBUG oslo_vmware.api [None req-62c1e732-39f8-407f-9e8d-e93f184a3c2b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333040, 'name': PowerOnVM_Task, 'duration_secs': 0.332255} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1409.677713] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-62c1e732-39f8-407f-9e8d-e93f184a3c2b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Powered on the VM {{(pid=62753) power_on_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1458}} [ 1409.677766] env[62753]: DEBUG nova.compute.manager [None req-62c1e732-39f8-407f-9e8d-e93f184a3c2b tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1409.678507] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-c138c9f3-b21a-4b49-97a1-9cdccff6489f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.348275] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-4cc6acb7-7a55-4c58-a562-42f74232165a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.355071] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-537d10ac-9af0-436b-9add-c78aa8adb4f7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Suspending the VM {{(pid=62753) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1162}} [ 1411.355341] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.SuspendVM_Task with opID=oslo.vmware-d5b98f4c-a012-4246-87cf-b60e0ceed622 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1411.362143] env[62753]: DEBUG oslo_vmware.api [None req-537d10ac-9af0-436b-9add-c78aa8adb4f7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1411.362143] env[62753]: value = "task-1333041" [ 1411.362143] env[62753]: _type = "Task" [ 1411.362143] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1411.371648] env[62753]: DEBUG oslo_vmware.api [None req-537d10ac-9af0-436b-9add-c78aa8adb4f7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333041, 'name': SuspendVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1411.873323] env[62753]: DEBUG oslo_vmware.api [None req-537d10ac-9af0-436b-9add-c78aa8adb4f7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333041, 'name': SuspendVM_Task} progress is 75%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1412.373163] env[62753]: DEBUG oslo_vmware.api [None req-537d10ac-9af0-436b-9add-c78aa8adb4f7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333041, 'name': SuspendVM_Task, 'duration_secs': 0.589924} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1412.373563] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-537d10ac-9af0-436b-9add-c78aa8adb4f7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Suspended the VM {{(pid=62753) suspend /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1166}} [ 1412.373622] env[62753]: DEBUG nova.compute.manager [None req-537d10ac-9af0-436b-9add-c78aa8adb4f7 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1412.374360] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f0a3803d-8886-4c25-a432-40ce43a3cea7 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1413.674935] env[62753]: INFO nova.compute.manager [None req-57ff5dfe-306f-4618-be8c-6bb6f8e8e7ae tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Resuming [ 1413.675623] env[62753]: DEBUG nova.objects.instance [None req-57ff5dfe-306f-4618-be8c-6bb6f8e8e7ae tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lazy-loading 'flavor' on Instance uuid a3053e3f-9dc7-4724-bd6d-4deaf4ee931c {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1414.684234] env[62753]: DEBUG oslo_concurrency.lockutils [None req-57ff5dfe-306f-4618-be8c-6bb6f8e8e7ae tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:310}} [ 1414.684649] env[62753]: DEBUG oslo_concurrency.lockutils [None req-57ff5dfe-306f-4618-be8c-6bb6f8e8e7ae tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquired lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:313}} [ 1414.684649] env[62753]: DEBUG nova.network.neutron [None req-57ff5dfe-306f-4618-be8c-6bb6f8e8e7ae tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Building network info cache for instance {{(pid=62753) _get_instance_nw_info /opt/stack/nova/nova/network/neutron.py:2067}} [ 1415.383779] env[62753]: DEBUG nova.network.neutron [None req-57ff5dfe-306f-4618-be8c-6bb6f8e8e7ae tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Updating instance_info_cache with network_info: [{"id": "176523fa-3534-4c71-a62e-d8ce87b429de", "address": "fa:16:3e:90:9d:46", "network": {"id": "bb29ec92-3bfe-43c9-9671-fd99901ce6a7", "bridge": "br-int", "label": "tempest-ServerActionsTestJSON-955549118-network", "subnets": [{"cidr": "192.168.128.0/28", "dns": [], "gateway": {"address": "192.168.128.1", "type": "gateway", "version": 4, "meta": {}}, "ips": [{"address": "192.168.128.10", "type": "fixed", "version": 4, "meta": {}, "floating_ips": [{"address": "10.180.180.196", "type": "floating", "version": 4, "meta": {}}]}], "routes": [], "version": 4, "meta": {"enable_dhcp": true, "dhcp_server": "192.168.128.2"}}], "meta": {"injected": false, "tenant_id": "891d1efbcac34ceba5fa72ea6b5fe38b", "mtu": 8950, "physical_network": "default", "tunneled": false}}, "type": "ovs", "details": {"connectivity": "l2", "port_filter": true, "nsx-logical-switch-id": "cae1d6a8-cbba-4bbf-af10-ba5467340475", "external-id": "nsx-vlan-transportzone-271", "segmentation_id": 271, "bound_drivers": {"0": "nsxv3"}}, "devname": "tap176523fa-35", "ovs_interfaceid": "176523fa-3534-4c71-a62e-d8ce87b429de", "qbh_params": null, "qbg_params": null, "active": true, "vnic_type": "normal", "profile": {}, "preserve_on_delete": false, "delegate_create": true, "meta": {}}] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1415.887082] env[62753]: DEBUG oslo_concurrency.lockutils [None req-57ff5dfe-306f-4618-be8c-6bb6f8e8e7ae tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Releasing lock "refresh_cache-a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" {{(pid=62753) lock /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:331}} [ 1415.888083] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-2be79110-a6de-4324-903b-0380fa2fb9c1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.895054] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-57ff5dfe-306f-4618-be8c-6bb6f8e8e7ae tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Resuming the VM {{(pid=62753) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1183}} [ 1415.895276] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOnVM_Task with opID=oslo.vmware-6395555f-5216-4647-ad92-c3566a22526b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1415.900910] env[62753]: DEBUG oslo_vmware.api [None req-57ff5dfe-306f-4618-be8c-6bb6f8e8e7ae tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1415.900910] env[62753]: value = "task-1333042" [ 1415.900910] env[62753]: _type = "Task" [ 1415.900910] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1415.908076] env[62753]: DEBUG oslo_vmware.api [None req-57ff5dfe-306f-4618-be8c-6bb6f8e8e7ae tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333042, 'name': PowerOnVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1416.410258] env[62753]: DEBUG oslo_vmware.api [None req-57ff5dfe-306f-4618-be8c-6bb6f8e8e7ae tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333042, 'name': PowerOnVM_Task, 'duration_secs': 0.482505} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1416.410549] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-57ff5dfe-306f-4618-be8c-6bb6f8e8e7ae tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Resumed the VM {{(pid=62753) resume /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1188}} [ 1416.410735] env[62753]: DEBUG nova.compute.manager [None req-57ff5dfe-306f-4618-be8c-6bb6f8e8e7ae tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Checking state {{(pid=62753) _get_power_state /opt/stack/nova/nova/compute/manager.py:1791}} [ 1416.411560] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-83f965d2-91ba-476e-83aa-019cf759e985 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.262782] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1417.263237] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" acquired by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: waited 0.001s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1417.264774] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "a3053e3f-9dc7-4724-bd6d-4deaf4ee931c-events" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1417.265010] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "a3053e3f-9dc7-4724-bd6d-4deaf4ee931c-events" acquired by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: waited 0.002s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1417.265209] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "a3053e3f-9dc7-4724-bd6d-4deaf4ee931c-events" "released" by "nova.compute.manager.InstanceEvents.clear_events_for_instance.._clear_events" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1417.267426] env[62753]: INFO nova.compute.manager [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Terminating instance [ 1417.269369] env[62753]: DEBUG nova.compute.manager [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Start destroying the instance on the hypervisor. {{(pid=62753) _shutdown_instance /opt/stack/nova/nova/compute/manager.py:3150}} [ 1417.269568] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Destroying instance {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1141}} [ 1417.270439] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-da9aaf4e-c132-491a-9e26-dc42567623d1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.278527] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Powering off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1512}} [ 1417.278806] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.PowerOffVM_Task with opID=oslo.vmware-54383b05-f2d9-4c99-964a-13892d856322 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1417.285234] env[62753]: DEBUG oslo_vmware.api [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1417.285234] env[62753]: value = "task-1333043" [ 1417.285234] env[62753]: _type = "Task" [ 1417.285234] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1417.293458] env[62753]: DEBUG oslo_vmware.api [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333043, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1417.796188] env[62753]: DEBUG oslo_vmware.api [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333043, 'name': PowerOffVM_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1418.296554] env[62753]: DEBUG oslo_vmware.api [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333043, 'name': PowerOffVM_Task, 'duration_secs': 0.765165} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1418.296944] env[62753]: DEBUG nova.virt.vmwareapi.vm_util [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Powered off the VM {{(pid=62753) power_off_instance /opt/stack/nova/nova/virt/vmwareapi/vm_util.py:1517}} [ 1418.297057] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Unregistering the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1093}} [ 1418.297309] env[62753]: DEBUG oslo_vmware.service [-] Invoking VirtualMachine.UnregisterVM with opID=oslo.vmware-dc513706-3d96-4132-acc8-c8db1a1fa917 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.439030] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Unregistered the VM {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1096}} [ 1418.439301] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Deleting contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1106}} [ 1418.439500] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Deleting the datastore file [datastore2] a3053e3f-9dc7-4724-bd6d-4deaf4ee931c {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:211}} [ 1418.439824] env[62753]: DEBUG oslo_vmware.service [-] Invoking FileManager.DeleteDatastoreFile_Task with opID=oslo.vmware-44ff5780-9f09-4b48-b4b9-f0efb3366ac2 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1418.445712] env[62753]: DEBUG oslo_vmware.api [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for the task: (returnval){ [ 1418.445712] env[62753]: value = "task-1333045" [ 1418.445712] env[62753]: _type = "Task" [ 1418.445712] env[62753]: } to complete. {{(pid=62753) wait_for_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:397}} [ 1418.453188] env[62753]: DEBUG oslo_vmware.api [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333045, 'name': DeleteDatastoreFile_Task} progress is 0%. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:434}} [ 1418.955931] env[62753]: DEBUG oslo_vmware.api [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Task: {'id': task-1333045, 'name': DeleteDatastoreFile_Task, 'duration_secs': 0.146471} completed successfully. {{(pid=62753) _poll_task /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/api.py:444}} [ 1418.956200] env[62753]: DEBUG nova.virt.vmwareapi.ds_util [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Deleted the datastore file {{(pid=62753) file_delete /opt/stack/nova/nova/virt/vmwareapi/ds_util.py:220}} [ 1418.956382] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Deleted contents of the VM from datastore datastore2 {{(pid=62753) _destroy_instance /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1116}} [ 1418.956564] env[62753]: DEBUG nova.virt.vmwareapi.vmops [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Instance destroyed {{(pid=62753) destroy /opt/stack/nova/nova/virt/vmwareapi/vmops.py:1143}} [ 1418.956744] env[62753]: INFO nova.compute.manager [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Took 1.69 seconds to destroy the instance on the hypervisor. [ 1418.956988] env[62753]: DEBUG oslo.service.loopingcall [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Waiting for function nova.compute.manager.ComputeManager._try_deallocate_network.._deallocate_network_with_retries to return. {{(pid=62753) func /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/loopingcall.py:435}} [ 1418.957234] env[62753]: DEBUG nova.compute.manager [-] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Deallocating network for instance {{(pid=62753) _deallocate_network /opt/stack/nova/nova/compute/manager.py:2289}} [ 1418.957356] env[62753]: DEBUG nova.network.neutron [-] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] deallocate_for_instance() {{(pid=62753) deallocate_for_instance /opt/stack/nova/nova/network/neutron.py:1860}} [ 1419.383994] env[62753]: DEBUG nova.compute.manager [req-e662268d-3f5a-4f1d-a421-d60177bc9f80 req-9982be10-7978-453e-b991-6be078201ebd service nova] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Received event network-vif-deleted-176523fa-3534-4c71-a62e-d8ce87b429de {{(pid=62753) external_instance_event /opt/stack/nova/nova/compute/manager.py:11131}} [ 1419.383994] env[62753]: INFO nova.compute.manager [req-e662268d-3f5a-4f1d-a421-d60177bc9f80 req-9982be10-7978-453e-b991-6be078201ebd service nova] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Neutron deleted interface 176523fa-3534-4c71-a62e-d8ce87b429de; detaching it from the instance and deleting it from the info cache [ 1419.384448] env[62753]: DEBUG nova.network.neutron [req-e662268d-3f5a-4f1d-a421-d60177bc9f80 req-9982be10-7978-453e-b991-6be078201ebd service nova] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1419.864455] env[62753]: DEBUG nova.network.neutron [-] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Updating instance_info_cache with network_info: [] {{(pid=62753) update_instance_cache_with_nw_info /opt/stack/nova/nova/network/neutron.py:116}} [ 1419.886759] env[62753]: DEBUG oslo_vmware.service [-] Invoking SearchIndex.FindAllByUuid with opID=oslo.vmware-240486a1-3624-489b-8a81-c6704ce5259e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.896021] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-640b0084-8c8d-4dc0-a7ca-0962a5087823 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1419.917049] env[62753]: DEBUG nova.compute.manager [req-e662268d-3f5a-4f1d-a421-d60177bc9f80 req-9982be10-7978-453e-b991-6be078201ebd service nova] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Detach interface failed, port_id=176523fa-3534-4c71-a62e-d8ce87b429de, reason: Instance a3053e3f-9dc7-4724-bd6d-4deaf4ee931c could not be found. {{(pid=62753) _process_instance_vif_deleted_event /opt/stack/nova/nova/compute/manager.py:10965}} [ 1420.367228] env[62753]: INFO nova.compute.manager [-] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Took 1.41 seconds to deallocate network for instance. [ 1420.873746] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.update_usage" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1420.874155] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1420.874267] env[62753]: DEBUG nova.objects.instance [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lazy-loading 'resources' on Instance uuid a3053e3f-9dc7-4724-bd6d-4deaf4ee931c {{(pid=62753) obj_load_attr /opt/stack/nova/nova/objects/instance.py:1141}} [ 1421.409310] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-40250083-05e0-4195-a934-8fa54259cbc8 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.418417] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-f335ddf3-57f2-4cf3-906b-5d14cf7ff3f3 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.447470] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-8ae21de2-7a5a-4368-8eff-01c6366b900a {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.454233] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-cdd69fb7-a8c7-4d2b-8900-cd451f67b338 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1421.466708] env[62753]: DEBUG nova.compute.provider_tree [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1421.969643] env[62753]: DEBUG nova.scheduler.client.report [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1422.475274] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.update_usage" :: held 1.601s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1422.495712] env[62753]: INFO nova.scheduler.client.report [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Deleted allocations for instance a3053e3f-9dc7-4724-bd6d-4deaf4ee931c [ 1423.003953] env[62753]: DEBUG oslo_concurrency.lockutils [None req-b6dfffc3-6a54-4866-b1c6-9f12cab79af1 tempest-ServerActionsTestJSON-1900367229 tempest-ServerActionsTestJSON-1900367229-project-member] Lock "a3053e3f-9dc7-4724-bd6d-4deaf4ee931c" "released" by "nova.compute.manager.ComputeManager.terminate_instance..do_terminate_instance" :: held 5.741s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1425.010460] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._cleanup_expired_console_auth_tokens {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1430.512632] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._reclaim_queued_deletes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1430.513036] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] CONF.reclaim_instance_interval <= 0, skipping... {{(pid=62753) _reclaim_queued_deletes /opt/stack/nova/nova/compute/manager.py:10547}} [ 1431.010654] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_volume_usage {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1432.010013] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._instance_usage_audit {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1433.010457] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_rescued_instances {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1433.010833] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._poll_unconfirmed_resizes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1435.005691] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._check_instance_build_time {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1435.010364] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager.update_available_resource {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1435.513924] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1435.514183] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1435.514349] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker.clean_compute_node_cache" :: held 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1435.514506] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Auditing locally available compute resources for cpu-1 (node: domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28) {{(pid=62753) update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:907}} [ 1435.515420] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-d135d720-d004-4389-8bc2-1c2d12fd9f3e {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.523753] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-99a7b345-fbdf-4993-82c8-ba80790082ed {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.537421] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-22e9cce1-b30f-4e71-917a-0241fd22431b {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.543516] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-1e6ddc13-fd93-487c-ac66-a97a3ad9f8ad {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1435.570945] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Hypervisor/Node resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 free_ram=181286MB free_disk=131GB free_vcpus=48 pci_devices=None {{(pid=62753) _report_hypervisor_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1106}} [ 1435.571099] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Acquiring lock "compute_resources" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:402}} [ 1435.571293] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" acquired by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: waited 0.000s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:407}} [ 1436.700750] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Total usable vcpus: 48, total allocated vcpus: 0 {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1129}} [ 1436.700985] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Final resource view: name=domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 phys_ram=196590MB used_ram=512MB phys_disk=200GB used_disk=0GB total_vcpus=48 used_vcpus=0 pci_stats=[] {{(pid=62753) _report_final_resource_view /opt/stack/nova/nova/compute/resource_tracker.py:1138}} [ 1436.717276] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Refreshing inventories for resource provider 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:818}} [ 1436.729672] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Updating ProviderTree inventory for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 from _refresh_and_get_inventory using data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) _refresh_and_get_inventory /opt/stack/nova/nova/scheduler/client/report.py:782}} [ 1436.729858] env[62753]: DEBUG nova.compute.provider_tree [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Updating inventory in ProviderTree for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 with inventory: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:176}} [ 1436.739698] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Refreshing aggregate associations for resource provider 42c8603e-76bc-4e29-ad77-cac0094f7308, aggregates: None {{(pid=62753) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:827}} [ 1436.755470] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Refreshing trait associations for resource provider 42c8603e-76bc-4e29-ad77-cac0094f7308, traits: COMPUTE_NET_ATTACH_INTERFACE,HW_ARCH_X86_64,COMPUTE_IMAGE_TYPE_ISO,COMPUTE_NODE,COMPUTE_IMAGE_TYPE_VMDK,COMPUTE_SAME_HOST_COLD_MIGRATE {{(pid=62753) _refresh_associations /opt/stack/nova/nova/scheduler/client/report.py:839}} [ 1436.766939] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-3882b82e-8cbb-4540-94eb-69c34a9817a1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.774724] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-054c070a-91ce-4e83-8494-7cb5f0ab01b4 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.804512] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-a039129a-7ee1-4a4f-b8a6-4dd9dcbe9d0f {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.811071] env[62753]: DEBUG oslo_vmware.service [-] Invoking PropertyCollector.RetrievePropertiesEx with opID=oslo.vmware-acb3c097-8e49-424f-a80c-5be9583ce1b1 {{(pid=62753) request_handler /opt/stack/data/venv/lib/python3.10/site-packages/oslo_vmware/service.py:371}} [ 1436.823410] env[62753]: DEBUG nova.compute.provider_tree [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed in ProviderTree for provider: 42c8603e-76bc-4e29-ad77-cac0094f7308 {{(pid=62753) update_inventory /opt/stack/nova/nova/compute/provider_tree.py:180}} [ 1437.326800] env[62753]: DEBUG nova.scheduler.client.report [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Inventory has not changed for provider 42c8603e-76bc-4e29-ad77-cac0094f7308 based on inventory data: {'VCPU': {'total': 48, 'reserved': 0, 'min_unit': 1, 'max_unit': 16, 'step_size': 1, 'allocation_ratio': 4.0}, 'MEMORY_MB': {'total': 196590, 'reserved': 512, 'min_unit': 1, 'max_unit': 65530, 'step_size': 1, 'allocation_ratio': 1.0}, 'DISK_GB': {'total': 400, 'reserved': 0, 'min_unit': 1, 'max_unit': 131, 'step_size': 1, 'allocation_ratio': 1.0}} {{(pid=62753) set_inventory_for_provider /opt/stack/nova/nova/scheduler/client/report.py:954}} [ 1437.831736] env[62753]: DEBUG nova.compute.resource_tracker [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Compute_service record updated for cpu-1:domain-c8.fc996f14-c53b-4953-92e3-bdfa48f5cc28 {{(pid=62753) _update_available_resource /opt/stack/nova/nova/compute/resource_tracker.py:1067}} [ 1437.831736] env[62753]: DEBUG oslo_concurrency.lockutils [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Lock "compute_resources" "released" by "nova.compute.resource_tracker.ResourceTracker._update_available_resource" :: held 2.260s {{(pid=62753) inner /opt/stack/data/venv/lib/python3.10/site-packages/oslo_concurrency/lockutils.py:421}} [ 1439.010456] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._heal_instance_info_cache {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1439.010858] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Starting heal instance info cache {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9928}} [ 1439.010858] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Rebuilding the list of instances to heal {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:9932}} [ 1439.514080] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Didn't find any instances for network info cache update. {{(pid=62753) _heal_instance_info_cache /opt/stack/nova/nova/compute/manager.py:10014}} [ 1439.514362] env[62753]: DEBUG oslo_service.periodic_task [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Running periodic task ComputeManager._run_pending_deletes {{(pid=62753) run_periodic_tasks /opt/stack/data/venv/lib/python3.10/site-packages/oslo_service/periodic_task.py:210}} [ 1439.514518] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] Cleaning up deleted instances {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11228}} [ 1440.021453] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] There are 12 instances to clean {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11237}} [ 1440.021768] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: a3053e3f-9dc7-4724-bd6d-4deaf4ee931c] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1440.524552] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 5b196309-161a-4770-90c8-81b57909cb12] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1441.028841] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 40c28133-7363-4a8e-bd77-fce3bbfbee69] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1441.532556] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 221f76f0-ee69-4119-b19b-c148e40c1126] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1442.037501] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 2e784985-7045-448e-a650-f6692bb69d94] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1442.539058] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 6a97171d-7ea2-4581-8707-b6be33f1e8ed] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1443.041987] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 21e6cf03-e528-4289-8813-86020a4e8d2a] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1443.545353] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: ec0b07b8-1cb9-4911-8b03-efb466a21888] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1444.048543] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 51790f51-c464-48cc-9891-41cae5a8f63e] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1444.552569] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: a2e7fa98-566e-40f1-8da7-9318de89c14c] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1445.056336] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: 22cea165-cfde-403f-a0e2-82a86f5baa51] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}} [ 1445.560090] env[62753]: DEBUG nova.compute.manager [None req-687c3a17-3a9c-49bc-a3bc-b2cc01b59f6a None None] [instance: ca01cd3f-7f0e-4c61-a704-74f9945792b4] Instance has had 0 of 5 cleanup attempts {{(pid=62753) _run_pending_deletes /opt/stack/nova/nova/compute/manager.py:11241}}